var/home/core/zuul-output/0000755000175000017500000000000015114511744014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114523743015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005132376515114523735017717 0ustar rootrootDec 05 08:15:34 crc systemd[1]: Starting Kubernetes Kubelet... Dec 05 08:15:34 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:34 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 08:15:35 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 05 08:15:35 crc kubenswrapper[4876]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 08:15:35 crc kubenswrapper[4876]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 05 08:15:35 crc kubenswrapper[4876]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 08:15:35 crc kubenswrapper[4876]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 08:15:35 crc kubenswrapper[4876]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 05 08:15:35 crc kubenswrapper[4876]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.632576 4876 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637490 4876 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637518 4876 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637527 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637556 4876 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637564 4876 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637572 4876 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637580 4876 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637588 4876 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637597 4876 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637605 4876 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637612 4876 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637620 4876 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637627 4876 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637635 4876 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637643 4876 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637651 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637666 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637674 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637682 4876 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637690 4876 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637700 4876 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637710 4876 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637718 4876 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637726 4876 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637734 4876 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637742 4876 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637750 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637761 4876 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637772 4876 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637781 4876 feature_gate.go:330] unrecognized feature gate: Example Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637792 4876 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637801 4876 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637810 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637817 4876 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637826 4876 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637833 4876 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637841 4876 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637849 4876 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637859 4876 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637867 4876 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637878 4876 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637888 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637923 4876 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637932 4876 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637944 4876 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637955 4876 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637964 4876 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637972 4876 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637980 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637988 4876 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.637996 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638004 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638012 4876 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638020 4876 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638028 4876 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638036 4876 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638043 4876 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638053 4876 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638060 4876 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638068 4876 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638076 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638083 4876 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638092 4876 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638100 4876 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638107 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638115 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638125 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638133 4876 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638140 4876 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638148 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.638155 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638548 4876 flags.go:64] FLAG: --address="0.0.0.0" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638577 4876 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638597 4876 flags.go:64] FLAG: --anonymous-auth="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638609 4876 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638621 4876 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638631 4876 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638643 4876 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638654 4876 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638663 4876 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638672 4876 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638682 4876 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638691 4876 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638700 4876 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638709 4876 flags.go:64] FLAG: --cgroup-root="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638719 4876 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638729 4876 flags.go:64] FLAG: --client-ca-file="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638738 4876 flags.go:64] FLAG: --cloud-config="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638746 4876 flags.go:64] FLAG: --cloud-provider="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638755 4876 flags.go:64] FLAG: --cluster-dns="[]" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638779 4876 flags.go:64] FLAG: --cluster-domain="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638788 4876 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638797 4876 flags.go:64] FLAG: --config-dir="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638806 4876 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638815 4876 flags.go:64] FLAG: --container-log-max-files="5" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638828 4876 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638837 4876 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638846 4876 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638856 4876 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638865 4876 flags.go:64] FLAG: --contention-profiling="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638874 4876 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638883 4876 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638894 4876 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638935 4876 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638948 4876 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638957 4876 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638966 4876 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638976 4876 flags.go:64] FLAG: --enable-load-reader="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638985 4876 flags.go:64] FLAG: --enable-server="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.638994 4876 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639012 4876 flags.go:64] FLAG: --event-burst="100" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639023 4876 flags.go:64] FLAG: --event-qps="50" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639032 4876 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639041 4876 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639050 4876 flags.go:64] FLAG: --eviction-hard="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639062 4876 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639071 4876 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639080 4876 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639089 4876 flags.go:64] FLAG: --eviction-soft="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639098 4876 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639107 4876 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639116 4876 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639125 4876 flags.go:64] FLAG: --experimental-mounter-path="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639134 4876 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639143 4876 flags.go:64] FLAG: --fail-swap-on="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639152 4876 flags.go:64] FLAG: --feature-gates="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639167 4876 flags.go:64] FLAG: --file-check-frequency="20s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639176 4876 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639186 4876 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639195 4876 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639205 4876 flags.go:64] FLAG: --healthz-port="10248" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639215 4876 flags.go:64] FLAG: --help="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639226 4876 flags.go:64] FLAG: --hostname-override="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639238 4876 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639250 4876 flags.go:64] FLAG: --http-check-frequency="20s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639262 4876 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639274 4876 flags.go:64] FLAG: --image-credential-provider-config="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639284 4876 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639295 4876 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639304 4876 flags.go:64] FLAG: --image-service-endpoint="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639313 4876 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639322 4876 flags.go:64] FLAG: --kube-api-burst="100" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639332 4876 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639343 4876 flags.go:64] FLAG: --kube-api-qps="50" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639352 4876 flags.go:64] FLAG: --kube-reserved="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639361 4876 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639370 4876 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639379 4876 flags.go:64] FLAG: --kubelet-cgroups="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639387 4876 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639397 4876 flags.go:64] FLAG: --lock-file="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639405 4876 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639414 4876 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639423 4876 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639438 4876 flags.go:64] FLAG: --log-json-split-stream="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639447 4876 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639456 4876 flags.go:64] FLAG: --log-text-split-stream="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639466 4876 flags.go:64] FLAG: --logging-format="text" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639474 4876 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639484 4876 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639493 4876 flags.go:64] FLAG: --manifest-url="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639501 4876 flags.go:64] FLAG: --manifest-url-header="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639513 4876 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639524 4876 flags.go:64] FLAG: --max-open-files="1000000" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639535 4876 flags.go:64] FLAG: --max-pods="110" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639544 4876 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639553 4876 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639566 4876 flags.go:64] FLAG: --memory-manager-policy="None" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639576 4876 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639585 4876 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639594 4876 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639605 4876 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639626 4876 flags.go:64] FLAG: --node-status-max-images="50" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639635 4876 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639644 4876 flags.go:64] FLAG: --oom-score-adj="-999" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639667 4876 flags.go:64] FLAG: --pod-cidr="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639677 4876 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639696 4876 flags.go:64] FLAG: --pod-manifest-path="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639705 4876 flags.go:64] FLAG: --pod-max-pids="-1" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639715 4876 flags.go:64] FLAG: --pods-per-core="0" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639725 4876 flags.go:64] FLAG: --port="10250" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639734 4876 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639743 4876 flags.go:64] FLAG: --provider-id="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639751 4876 flags.go:64] FLAG: --qos-reserved="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639760 4876 flags.go:64] FLAG: --read-only-port="10255" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639770 4876 flags.go:64] FLAG: --register-node="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639779 4876 flags.go:64] FLAG: --register-schedulable="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639788 4876 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639802 4876 flags.go:64] FLAG: --registry-burst="10" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639811 4876 flags.go:64] FLAG: --registry-qps="5" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639820 4876 flags.go:64] FLAG: --reserved-cpus="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639829 4876 flags.go:64] FLAG: --reserved-memory="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639840 4876 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639849 4876 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639858 4876 flags.go:64] FLAG: --rotate-certificates="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639867 4876 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639876 4876 flags.go:64] FLAG: --runonce="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639885 4876 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639894 4876 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639931 4876 flags.go:64] FLAG: --seccomp-default="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639944 4876 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639953 4876 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639962 4876 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639971 4876 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639981 4876 flags.go:64] FLAG: --storage-driver-password="root" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.639991 4876 flags.go:64] FLAG: --storage-driver-secure="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640001 4876 flags.go:64] FLAG: --storage-driver-table="stats" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640011 4876 flags.go:64] FLAG: --storage-driver-user="root" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640020 4876 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640029 4876 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640038 4876 flags.go:64] FLAG: --system-cgroups="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640062 4876 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640077 4876 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640086 4876 flags.go:64] FLAG: --tls-cert-file="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640095 4876 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640110 4876 flags.go:64] FLAG: --tls-min-version="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640120 4876 flags.go:64] FLAG: --tls-private-key-file="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640129 4876 flags.go:64] FLAG: --topology-manager-policy="none" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640138 4876 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640147 4876 flags.go:64] FLAG: --topology-manager-scope="container" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640156 4876 flags.go:64] FLAG: --v="2" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640174 4876 flags.go:64] FLAG: --version="false" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640186 4876 flags.go:64] FLAG: --vmodule="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640197 4876 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.640206 4876 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640488 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640502 4876 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640510 4876 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640519 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640528 4876 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640536 4876 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640544 4876 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640554 4876 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640563 4876 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640570 4876 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640578 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640586 4876 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640593 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640601 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640609 4876 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640616 4876 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640624 4876 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640631 4876 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640639 4876 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640648 4876 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640656 4876 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640663 4876 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640678 4876 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640686 4876 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640694 4876 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640701 4876 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640710 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640720 4876 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640730 4876 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640739 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640747 4876 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640755 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640763 4876 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640770 4876 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640778 4876 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640786 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640794 4876 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640802 4876 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640809 4876 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640820 4876 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640828 4876 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640836 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640844 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640851 4876 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640859 4876 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640867 4876 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640874 4876 feature_gate.go:330] unrecognized feature gate: Example Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640882 4876 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640890 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640925 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640934 4876 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640942 4876 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640949 4876 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640957 4876 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640968 4876 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640977 4876 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640984 4876 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.640992 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641001 4876 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641009 4876 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641017 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641024 4876 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641032 4876 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641040 4876 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641047 4876 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641058 4876 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641067 4876 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641077 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641088 4876 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641098 4876 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.641109 4876 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.641124 4876 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.648963 4876 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.649212 4876 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649295 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649302 4876 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649306 4876 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649310 4876 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649314 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649318 4876 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649322 4876 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649325 4876 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649329 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649333 4876 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649338 4876 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649343 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649347 4876 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649351 4876 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649354 4876 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649358 4876 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649361 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649365 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649368 4876 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649372 4876 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649375 4876 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649380 4876 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649385 4876 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649389 4876 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649393 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649397 4876 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649403 4876 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649407 4876 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649411 4876 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649415 4876 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649419 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649424 4876 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649428 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649432 4876 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649436 4876 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649440 4876 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649444 4876 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649449 4876 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649452 4876 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649456 4876 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649460 4876 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649463 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649467 4876 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649470 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649474 4876 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649478 4876 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649481 4876 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649485 4876 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649488 4876 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649515 4876 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649520 4876 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649523 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649527 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649530 4876 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649534 4876 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649537 4876 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649541 4876 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649544 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649548 4876 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649551 4876 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649555 4876 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649558 4876 feature_gate.go:330] unrecognized feature gate: Example Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649562 4876 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649568 4876 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649572 4876 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649576 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649579 4876 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649583 4876 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649587 4876 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649590 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649595 4876 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.649602 4876 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649716 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649723 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649727 4876 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649731 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649735 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649739 4876 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649742 4876 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649746 4876 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649749 4876 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649753 4876 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649757 4876 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649760 4876 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649763 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649767 4876 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649771 4876 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649774 4876 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649778 4876 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649781 4876 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649785 4876 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649789 4876 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649794 4876 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649800 4876 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649804 4876 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649808 4876 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649814 4876 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649819 4876 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649823 4876 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649827 4876 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649831 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649835 4876 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649839 4876 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649843 4876 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649846 4876 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649850 4876 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649855 4876 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649859 4876 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649863 4876 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649868 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649872 4876 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649877 4876 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649881 4876 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649886 4876 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649890 4876 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649912 4876 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649917 4876 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649924 4876 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649930 4876 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649935 4876 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649938 4876 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649942 4876 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649946 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649950 4876 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649953 4876 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649957 4876 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649960 4876 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649965 4876 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649969 4876 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649973 4876 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649976 4876 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649981 4876 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649985 4876 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649989 4876 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649993 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.649996 4876 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.650000 4876 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.650004 4876 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.650008 4876 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.650011 4876 feature_gate.go:330] unrecognized feature gate: Example Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.650015 4876 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.650018 4876 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.650022 4876 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.650029 4876 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.650362 4876 server.go:940] "Client rotation is on, will bootstrap in background" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.655800 4876 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.656005 4876 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.656982 4876 server.go:997] "Starting client certificate rotation" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.657026 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.657375 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-15 14:20:43.314083126 +0000 UTC Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.657511 4876 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 246h5m7.656576939s for next certificate rotation Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.663824 4876 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.666630 4876 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.676853 4876 log.go:25] "Validated CRI v1 runtime API" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.697117 4876 log.go:25] "Validated CRI v1 image API" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.699195 4876 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.703038 4876 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-05-08-11-14-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.703129 4876 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.729958 4876 manager.go:217] Machine: {Timestamp:2025-12-05 08:15:35.727792635 +0000 UTC m=+0.216457337 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4a836e23-e8c7-4003-b121-650a726282d3 BootID:5aa7b9b5-1049-47fd-9576-7435395774d5 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8d:4c:f4 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8d:4c:f4 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:66:39:cc Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e6:e0:00 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:45:77:4c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:95:cb:25 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:22:9e:e9:33:33:46 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3a:ae:ca:54:48:3b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.730272 4876 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.730628 4876 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.731328 4876 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.731535 4876 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.731600 4876 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.731784 4876 topology_manager.go:138] "Creating topology manager with none policy" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.731793 4876 container_manager_linux.go:303] "Creating device plugin manager" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.731973 4876 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.732005 4876 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.732305 4876 state_mem.go:36] "Initialized new in-memory state store" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.732510 4876 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.733092 4876 kubelet.go:418] "Attempting to sync node with API server" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.733114 4876 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.733136 4876 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.733150 4876 kubelet.go:324] "Adding apiserver pod source" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.733164 4876 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.737634 4876 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.738293 4876 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.738832 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.739075 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.738849 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.740319 4876 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.740309 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741225 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741299 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741316 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741331 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741356 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741370 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741385 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741409 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741427 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741442 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741485 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741500 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.741816 4876 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.742730 4876 server.go:1280] "Started kubelet" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.743630 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.743695 4876 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.743764 4876 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.745068 4876 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.745160 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.745242 4876 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.745306 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 21:39:08.514656006 +0000 UTC Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.745350 4876 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 493h23m32.769309154s for next certificate rotation Dec 05 08:15:35 crc systemd[1]: Started Kubernetes Kubelet. Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.746013 4876 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.746029 4876 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.746148 4876 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.746404 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.747741 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="200ms" Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.750599 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.750764 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.751556 4876 factory.go:55] Registering systemd factory Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.751586 4876 factory.go:221] Registration of the systemd container factory successfully Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.751824 4876 server.go:460] "Adding debug handlers to kubelet server" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.752143 4876 factory.go:153] Registering CRI-O factory Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.752200 4876 factory.go:221] Registration of the crio container factory successfully Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.752347 4876 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.752418 4876 factory.go:103] Registering Raw factory Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.752457 4876 manager.go:1196] Started watching for new ooms in manager Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.753979 4876 manager.go:319] Starting recovery of all containers Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.755501 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e43b1ff05f509 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 08:15:35.742665993 +0000 UTC m=+0.231330655,LastTimestamp:2025-12-05 08:15:35.742665993 +0000 UTC m=+0.231330655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768380 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768459 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768482 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768503 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768523 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768546 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768567 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768588 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768612 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768631 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768652 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768673 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768720 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768743 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768764 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768783 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768802 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768821 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768840 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768859 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768877 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768895 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768941 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768960 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.768980 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769000 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769022 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769041 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769133 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769154 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769174 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769192 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769212 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769260 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769324 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769346 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769366 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769387 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769408 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769426 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769445 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769463 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769509 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769528 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769547 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769564 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769583 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769601 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769618 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769637 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769657 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769675 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769704 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769724 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769744 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769798 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769821 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769839 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769859 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769880 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769924 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769945 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769962 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769981 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.769999 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770017 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770037 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770056 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770076 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770096 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770117 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770134 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770155 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770173 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770191 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770210 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770227 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770245 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770265 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770285 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770305 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770324 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770342 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770362 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770380 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770399 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770417 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770436 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770456 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770474 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770492 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770510 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770528 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770550 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770568 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770588 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770612 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770631 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770650 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770668 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770687 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770707 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770727 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770745 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770776 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770896 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770945 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770968 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.770991 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771012 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771035 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771054 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771075 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771095 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771114 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771133 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771154 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771172 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771190 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771208 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771229 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771252 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771270 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771288 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771307 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771325 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771346 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771368 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771392 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771411 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771430 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771448 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771468 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771488 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771508 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771527 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771546 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771564 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771582 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771601 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771618 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771637 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771655 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771675 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771694 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771712 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771731 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771751 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771769 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771787 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771807 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771825 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771845 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771864 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.771883 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772696 4876 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772735 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772758 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772779 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772800 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772820 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772839 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772858 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772875 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772893 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.772987 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773011 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773032 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773056 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773108 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773135 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773162 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773187 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773209 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773228 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773246 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773266 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773284 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773305 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773323 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773344 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773363 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773384 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773403 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773425 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773447 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773466 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773484 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773503 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773520 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773539 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773559 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773578 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773597 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773617 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773637 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773657 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773680 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773702 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773721 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773744 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773763 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773782 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773803 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773824 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773844 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773863 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773883 4876 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773930 4876 reconstruct.go:97] "Volume reconstruction finished" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.773944 4876 reconciler.go:26] "Reconciler: start to sync state" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.786453 4876 manager.go:324] Recovery completed Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.797711 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.800338 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.800413 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.800429 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.801250 4876 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.801268 4876 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.801317 4876 state_mem.go:36] "Initialized new in-memory state store" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.814491 4876 policy_none.go:49] "None policy: Start" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.817477 4876 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.817514 4876 state_mem.go:35] "Initializing new in-memory state store" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.818301 4876 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.822536 4876 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.822585 4876 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.822614 4876 kubelet.go:2335] "Starting kubelet main sync loop" Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.822675 4876 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 05 08:15:35 crc kubenswrapper[4876]: W1205 08:15:35.824389 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.824469 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.847360 4876 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.874479 4876 manager.go:334] "Starting Device Plugin manager" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.874559 4876 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.874577 4876 server.go:79] "Starting device plugin registration server" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.875256 4876 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.875280 4876 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.876148 4876 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.876246 4876 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.876257 4876 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.884057 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.925981 4876 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.926219 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.928193 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.928237 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.928250 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.928412 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.928986 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.929141 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.929257 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.929292 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.929306 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.929617 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.929834 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.929946 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.930687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.930719 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.930733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.930816 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.930882 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.930945 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.931419 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.931548 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.931751 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.932388 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.932438 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.932459 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.932744 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.932792 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.932805 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.934219 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.934247 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.934257 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.934434 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.934519 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.934553 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.935561 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.935584 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.935594 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.935624 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.935686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.935706 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.936171 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.936252 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.938020 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.938037 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.938047 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.948747 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="400ms" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.976545 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.977559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.977649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.977672 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.977734 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 08:15:35 crc kubenswrapper[4876]: E1205 08:15:35.978392 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.980784 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.980870 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981100 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981160 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981219 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981296 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981362 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981397 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981467 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981538 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981573 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981644 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981748 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981828 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 08:15:35 crc kubenswrapper[4876]: I1205 08:15:35.981866 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084095 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084240 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084296 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084340 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084366 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084471 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084392 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084540 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084563 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084581 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084619 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084674 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084683 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084706 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084726 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084736 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084763 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084781 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084797 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084808 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084830 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084846 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084868 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084878 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084939 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084960 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.084975 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.085021 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.085123 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.179332 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.181127 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.181229 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.181241 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.181280 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 08:15:36 crc kubenswrapper[4876]: E1205 08:15:36.181848 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.253806 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.265197 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: W1205 08:15:36.281665 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8d07041aafc1c108250fafe0f6ae22b176bd89007c00e898162e655fa15995af WatchSource:0}: Error finding container 8d07041aafc1c108250fafe0f6ae22b176bd89007c00e898162e655fa15995af: Status 404 returned error can't find the container with id 8d07041aafc1c108250fafe0f6ae22b176bd89007c00e898162e655fa15995af Dec 05 08:15:36 crc kubenswrapper[4876]: W1205 08:15:36.285775 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3f7b6517396b9cb4c343c5901b0ee7d4891c22d986d3fbd38a25fcca411f35e5 WatchSource:0}: Error finding container 3f7b6517396b9cb4c343c5901b0ee7d4891c22d986d3fbd38a25fcca411f35e5: Status 404 returned error can't find the container with id 3f7b6517396b9cb4c343c5901b0ee7d4891c22d986d3fbd38a25fcca411f35e5 Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.285822 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.293582 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: W1205 08:15:36.306455 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-5e5acc0ad61436b7d387ee4e1a13bf2fd35914b12c6109e9b10670ba985c6ed1 WatchSource:0}: Error finding container 5e5acc0ad61436b7d387ee4e1a13bf2fd35914b12c6109e9b10670ba985c6ed1: Status 404 returned error can't find the container with id 5e5acc0ad61436b7d387ee4e1a13bf2fd35914b12c6109e9b10670ba985c6ed1 Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.314971 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:36 crc kubenswrapper[4876]: W1205 08:15:36.336019 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b253efd9dc31cbacfb008be41dfce6a607c7f3eb47a57792ce7ad2eaf5a96fe0 WatchSource:0}: Error finding container b253efd9dc31cbacfb008be41dfce6a607c7f3eb47a57792ce7ad2eaf5a96fe0: Status 404 returned error can't find the container with id b253efd9dc31cbacfb008be41dfce6a607c7f3eb47a57792ce7ad2eaf5a96fe0 Dec 05 08:15:36 crc kubenswrapper[4876]: E1205 08:15:36.349646 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="800ms" Dec 05 08:15:36 crc kubenswrapper[4876]: W1205 08:15:36.576730 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:36 crc kubenswrapper[4876]: E1205 08:15:36.576839 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.582493 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.584915 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.584955 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.584968 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.584993 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 08:15:36 crc kubenswrapper[4876]: E1205 08:15:36.585419 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Dec 05 08:15:36 crc kubenswrapper[4876]: W1205 08:15:36.594807 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:36 crc kubenswrapper[4876]: E1205 08:15:36.595069 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 05 08:15:36 crc kubenswrapper[4876]: W1205 08:15:36.674423 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:36 crc kubenswrapper[4876]: E1205 08:15:36.675569 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.744598 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.830004 4876 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941" exitCode=0 Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.830089 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.830272 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b253efd9dc31cbacfb008be41dfce6a607c7f3eb47a57792ce7ad2eaf5a96fe0"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.830411 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.832185 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.832261 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.832297 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.833194 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.833259 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"42eaafdd4a4d42949384e10942ae0aaf23231f5d7fd32397953afd03fb73a205"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.834880 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab" exitCode=0 Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.834943 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.834980 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5e5acc0ad61436b7d387ee4e1a13bf2fd35914b12c6109e9b10670ba985c6ed1"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.835102 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.836230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.836285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.836300 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.837613 4876 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c2b8e9319386fcda55671c0a949f53e95351bdd4f277c099f887d6d4abf79a6c" exitCode=0 Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.837652 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c2b8e9319386fcda55671c0a949f53e95351bdd4f277c099f887d6d4abf79a6c"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.838141 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3f7b6517396b9cb4c343c5901b0ee7d4891c22d986d3fbd38a25fcca411f35e5"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.838380 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.838412 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.839426 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.839495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.839511 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.839759 4876 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252" exitCode=0 Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.839805 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.839831 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8d07041aafc1c108250fafe0f6ae22b176bd89007c00e898162e655fa15995af"} Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.839920 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.840342 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.840375 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.840388 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.840796 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.840857 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:36 crc kubenswrapper[4876]: I1205 08:15:36.840881 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:37 crc kubenswrapper[4876]: E1205 08:15:37.152049 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="1.6s" Dec 05 08:15:37 crc kubenswrapper[4876]: W1205 08:15:37.372146 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Dec 05 08:15:37 crc kubenswrapper[4876]: E1205 08:15:37.372261 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.385631 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.386695 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.386735 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.386749 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.386777 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 08:15:37 crc kubenswrapper[4876]: E1205 08:15:37.387164 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.846749 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.846835 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.846859 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.846770 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.848123 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.848173 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.848192 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.855338 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.855378 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.855391 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.855404 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.855414 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.855564 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.856673 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.856705 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.856715 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.858170 4876 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="00dbb0f2de77fde179114484a6675b591d001eaa4f3c0c63428d543dfad744c2" exitCode=0 Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.858255 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"00dbb0f2de77fde179114484a6675b591d001eaa4f3c0c63428d543dfad744c2"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.858500 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.859469 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.859516 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.859534 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.859630 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9a056e141e0c391539ffd94a814372c68901b5e7ea03c7c696c333c5fa320142"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.859741 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.861145 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.861183 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.861192 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.862861 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.862889 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.862917 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6"} Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.862993 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.863989 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.864009 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:37 crc kubenswrapper[4876]: I1205 08:15:37.864017 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.097083 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.868116 4876 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5ccf99b978ab66c4199bc62d5469ba06bcf74b005054c5f66014b49cd8d70064" exitCode=0 Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.868259 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.868775 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5ccf99b978ab66c4199bc62d5469ba06bcf74b005054c5f66014b49cd8d70064"} Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.868868 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.869171 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.869461 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.869487 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.869498 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.869649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.869693 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.869715 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.870091 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.870144 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.870169 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.987344 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.989374 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.989435 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.989453 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:38 crc kubenswrapper[4876]: I1205 08:15:38.989488 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.262223 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.876200 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5b38cf2f83268a49a012abc45d8a404211da0b6c8380e1074f6d1b7ee4726d4e"} Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.876267 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.876281 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c196d9321c1feed99a6e29056dd20cff0865f468e1843baed0dcb9d2f7d165bd"} Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.876307 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fde96790bbc7c5165716a79d8dfb3640293637fadb463f25cb83ca2bf695c99d"} Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.876383 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.877600 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.877641 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.877656 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.877715 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.877748 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:39 crc kubenswrapper[4876]: I1205 08:15:39.877767 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.399229 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.886119 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b789ac89b1b87a2a77f2ebeda6aa6509b7a7d393c28bebab1efcb6d906810514"} Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.886263 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.886268 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6a85bf9c179161e5605536ac7afb43a0dc2445932c97d4d2e83b17df203b2d8c"} Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.886219 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.887966 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.888088 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.888153 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.888121 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.888287 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:40 crc kubenswrapper[4876]: I1205 08:15:40.888312 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.854956 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.889839 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.889946 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.891690 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.891815 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.891916 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.891968 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.892062 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:41 crc kubenswrapper[4876]: I1205 08:15:41.892093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.223723 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.224492 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.225754 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.225782 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.225791 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.230862 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.895257 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.896453 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.896487 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.896500 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.974936 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.975116 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.976353 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.976404 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:43 crc kubenswrapper[4876]: I1205 08:15:43.976419 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:45 crc kubenswrapper[4876]: I1205 08:15:45.615850 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:15:45 crc kubenswrapper[4876]: I1205 08:15:45.616085 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:45 crc kubenswrapper[4876]: I1205 08:15:45.617216 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:45 crc kubenswrapper[4876]: I1205 08:15:45.617269 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:45 crc kubenswrapper[4876]: I1205 08:15:45.617285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:45 crc kubenswrapper[4876]: E1205 08:15:45.884188 4876 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.327232 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.327625 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.329734 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.329780 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.329789 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.332162 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.906485 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.908085 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.908149 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:46 crc kubenswrapper[4876]: I1205 08:15:46.908169 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:47 crc kubenswrapper[4876]: I1205 08:15:47.720212 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:47 crc kubenswrapper[4876]: I1205 08:15:47.745511 4876 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 05 08:15:47 crc kubenswrapper[4876]: I1205 08:15:47.910357 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:47 crc kubenswrapper[4876]: I1205 08:15:47.911660 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:47 crc kubenswrapper[4876]: I1205 08:15:47.911737 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:47 crc kubenswrapper[4876]: I1205 08:15:47.911761 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:48 crc kubenswrapper[4876]: I1205 08:15:48.098669 4876 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 05 08:15:48 crc kubenswrapper[4876]: I1205 08:15:48.098754 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 05 08:15:48 crc kubenswrapper[4876]: W1205 08:15:48.576642 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 08:15:48 crc kubenswrapper[4876]: I1205 08:15:48.576773 4876 trace.go:236] Trace[1801961786]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 08:15:38.575) (total time: 10001ms): Dec 05 08:15:48 crc kubenswrapper[4876]: Trace[1801961786]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:15:48.576) Dec 05 08:15:48 crc kubenswrapper[4876]: Trace[1801961786]: [10.001281866s] [10.001281866s] END Dec 05 08:15:48 crc kubenswrapper[4876]: E1205 08:15:48.576808 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 08:15:48 crc kubenswrapper[4876]: W1205 08:15:48.745659 4876 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 08:15:48 crc kubenswrapper[4876]: I1205 08:15:48.745767 4876 trace.go:236] Trace[587466740]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 08:15:38.744) (total time: 10000ms): Dec 05 08:15:48 crc kubenswrapper[4876]: Trace[587466740]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (08:15:48.745) Dec 05 08:15:48 crc kubenswrapper[4876]: Trace[587466740]: [10.000891393s] [10.000891393s] END Dec 05 08:15:48 crc kubenswrapper[4876]: E1205 08:15:48.745792 4876 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 08:15:48 crc kubenswrapper[4876]: E1205 08:15:48.753760 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 05 08:15:48 crc kubenswrapper[4876]: I1205 08:15:48.940927 4876 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 08:15:48 crc kubenswrapper[4876]: I1205 08:15:48.940991 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 08:15:48 crc kubenswrapper[4876]: I1205 08:15:48.950363 4876 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 08:15:48 crc kubenswrapper[4876]: I1205 08:15:48.950430 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.104927 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.105096 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.106032 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.106070 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.106079 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.136505 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.327438 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.327544 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.915861 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.917085 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.917128 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.917139 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:49 crc kubenswrapper[4876]: I1205 08:15:49.948712 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 05 08:15:50 crc kubenswrapper[4876]: I1205 08:15:50.919183 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:50 crc kubenswrapper[4876]: I1205 08:15:50.921214 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:50 crc kubenswrapper[4876]: I1205 08:15:50.921275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:50 crc kubenswrapper[4876]: I1205 08:15:50.921294 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.863358 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.863585 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.865105 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.865147 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.865160 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.870611 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.921649 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.923226 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.923317 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:51 crc kubenswrapper[4876]: I1205 08:15:51.923345 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:53 crc kubenswrapper[4876]: I1205 08:15:53.934024 4876 trace.go:236] Trace[1359864970]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 08:15:40.347) (total time: 13586ms): Dec 05 08:15:53 crc kubenswrapper[4876]: Trace[1359864970]: ---"Objects listed" error: 13586ms (08:15:53.933) Dec 05 08:15:53 crc kubenswrapper[4876]: Trace[1359864970]: [13.586114721s] [13.586114721s] END Dec 05 08:15:53 crc kubenswrapper[4876]: I1205 08:15:53.934070 4876 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 08:15:53 crc kubenswrapper[4876]: I1205 08:15:53.935039 4876 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 05 08:15:53 crc kubenswrapper[4876]: E1205 08:15:53.937041 4876 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 05 08:15:53 crc kubenswrapper[4876]: I1205 08:15:53.938511 4876 trace.go:236] Trace[369653089]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 08:15:39.619) (total time: 14318ms): Dec 05 08:15:53 crc kubenswrapper[4876]: Trace[369653089]: ---"Objects listed" error: 14318ms (08:15:53.938) Dec 05 08:15:53 crc kubenswrapper[4876]: Trace[369653089]: [14.318574873s] [14.318574873s] END Dec 05 08:15:53 crc kubenswrapper[4876]: I1205 08:15:53.938542 4876 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.144023 4876 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51932->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.144078 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51932->192.168.126.11:17697: read: connection reset by peer" Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.144892 4876 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51942->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.144979 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51942->192.168.126.11:17697: read: connection reset by peer" Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.145330 4876 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.145359 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.640566 4876 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.881506 4876 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.932591 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.934032 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67" exitCode=255 Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.934063 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67"} Dec 05 08:15:54 crc kubenswrapper[4876]: I1205 08:15:54.944800 4876 scope.go:117] "RemoveContainer" containerID="ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.745029 4876 apiserver.go:52] "Watching apiserver" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.748668 4876 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.748959 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-kube-apiserver/kube-apiserver-crc","openshift-multus/multus-additional-cni-plugins-z44lm","openshift-multus/multus-jwrfx","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-ovn-kubernetes/ovnkube-node-58npg","openshift-dns/node-resolver-n4dsp","openshift-machine-config-operator/machine-config-daemon-xj8qx","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.749239 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.749294 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.749324 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.749382 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.749591 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.749819 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.749934 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.749981 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.750020 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.750334 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.750682 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.750747 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.750990 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.751237 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-n4dsp" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.754391 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.754625 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.754646 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.755177 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.755180 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.755241 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.757203 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.758081 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.758354 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.758527 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.758684 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.758867 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.759266 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.759525 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.759717 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.759894 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.760107 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.760333 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.760630 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.763576 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.763577 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.763694 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.763924 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.763951 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.763955 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.764202 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.764235 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.764251 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.764284 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.764361 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.764487 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.781185 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.798281 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.813048 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.823364 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.838015 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.845181 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.847986 4876 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.858663 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.868530 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.876799 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.886972 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.908328 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.922287 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.935342 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.937842 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.939578 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96"} Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.940609 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946092 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946141 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946167 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946193 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946217 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946237 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946261 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946285 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946309 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946329 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946348 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946368 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946381 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946388 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946433 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946450 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946466 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946462 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946481 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946543 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946563 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946580 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946599 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946618 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946635 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946682 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946697 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946712 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946727 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946741 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946756 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946774 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946804 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946820 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946834 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946851 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946859 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946869 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946886 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946884 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946928 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946953 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946975 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.946991 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947005 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947019 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947022 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947054 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947070 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947087 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947099 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947112 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947104 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947158 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947162 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947187 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947214 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947239 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947265 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947278 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947291 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947316 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947340 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947349 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947364 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947389 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947420 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947442 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947466 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947490 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947514 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947536 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947557 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947577 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947597 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947619 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947644 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947667 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947687 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947711 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947735 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947755 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947777 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947798 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947820 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947841 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947861 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947882 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947952 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947975 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947998 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948023 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948048 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948080 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948102 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948126 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948149 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948173 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948197 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948222 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948245 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948271 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948296 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948319 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948347 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948373 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948400 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948423 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948446 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948467 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948488 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948511 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948533 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948555 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948576 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948595 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948624 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948645 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948666 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948687 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948708 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948728 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948748 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948771 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948796 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948820 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948846 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948868 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948889 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948933 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948954 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948975 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948997 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949019 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949040 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949066 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949086 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949112 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949135 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949158 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949192 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949214 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949310 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949336 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949359 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949383 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949404 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949426 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949449 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949470 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949492 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949514 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949535 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949559 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949582 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949608 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949636 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949658 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949682 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949733 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949762 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949785 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949807 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949832 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949858 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949882 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950108 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950137 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950160 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950183 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950207 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950232 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950257 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950280 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950303 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950325 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950348 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950369 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950391 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950414 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950436 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950458 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950480 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950503 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950525 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950550 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950575 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950597 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950621 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950645 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950667 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950690 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950713 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950736 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950760 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950784 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950806 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950830 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950853 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950875 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950919 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950945 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950967 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.950990 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951014 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951077 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951110 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951135 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cnibin\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951159 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e7274e0d-7f30-495c-800a-5dcbf1656b0d-cni-binary-copy\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951181 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-slash\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951202 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-netd\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951230 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951254 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h79nh\" (UniqueName: \"kubernetes.io/projected/905a4c7a-1ee3-46ff-bd6b-04a34078015c-kube-api-access-h79nh\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951274 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-system-cni-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951454 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-cnibin\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951479 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-os-release\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951499 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-k8s-cni-cncf-io\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951520 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-netns\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951543 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-multus-certs\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951564 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-ovn-kubernetes\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951588 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-socket-dir-parent\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951614 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951641 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951664 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-os-release\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951687 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-cni-multus\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951714 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/77322cc8-c6ab-4250-8098-9938309f0af8-rootfs\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951737 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.951765 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-daemon-config\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952284 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-node-log\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952314 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-etc-kubernetes\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952337 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/446e4833-20af-47c0-b4e0-7116c3972366-ovn-node-metrics-cert\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952362 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-cni-bin\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952383 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952405 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-config\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952431 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952458 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952482 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9djw\" (UniqueName: \"kubernetes.io/projected/77322cc8-c6ab-4250-8098-9938309f0af8-kube-api-access-v9djw\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952503 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98p7w\" (UniqueName: \"kubernetes.io/projected/e7274e0d-7f30-495c-800a-5dcbf1656b0d-kube-api-access-98p7w\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952530 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952552 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwhkw\" (UniqueName: \"kubernetes.io/projected/ae897c0d-b078-40d3-bbd4-aa2c279c89ec-kube-api-access-fwhkw\") pod \"node-resolver-n4dsp\" (UID: \"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\") " pod="openshift-dns/node-resolver-n4dsp" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952576 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/77322cc8-c6ab-4250-8098-9938309f0af8-mcd-auth-proxy-config\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952595 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-hostroot\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952613 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-systemd-units\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952633 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-systemd\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952655 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-log-socket\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952677 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-env-overrides\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952698 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-etc-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952719 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ae897c0d-b078-40d3-bbd4-aa2c279c89ec-hosts-file\") pod \"node-resolver-n4dsp\" (UID: \"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\") " pod="openshift-dns/node-resolver-n4dsp" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952748 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952772 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-bin\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952798 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952820 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952843 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-conf-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952864 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-var-lib-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952885 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952925 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/77322cc8-c6ab-4250-8098-9938309f0af8-proxy-tls\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952953 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952979 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cni-binary-copy\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953003 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953029 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953058 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953085 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-kubelet\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953109 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-kubelet\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953133 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcg2f\" (UniqueName: \"kubernetes.io/projected/446e4833-20af-47c0-b4e0-7116c3972366-kube-api-access-fcg2f\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953154 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-script-lib\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953178 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-system-cni-dir\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953203 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-cni-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953226 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-netns\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953248 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-ovn\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953319 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953337 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953351 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953366 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953381 4876 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953397 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953411 4876 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953432 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953448 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953461 4876 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.952299 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961347 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.962874 4876 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947515 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.969399 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947590 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.947644 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948020 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948139 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948275 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948316 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948327 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948483 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948515 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948521 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948669 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.948720 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.949030 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953096 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953167 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953243 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953306 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953460 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953608 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953883 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953923 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953925 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.953975 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.954080 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.954135 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.954361 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.954363 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.954602 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.954728 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.955076 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.955204 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.955728 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.955813 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.956580 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.956938 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.957010 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.957139 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:15:56.457090056 +0000 UTC m=+20.945754678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.957206 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.957212 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.957539 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.957615 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.958115 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.958478 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.958644 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.958786 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.959016 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.959105 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.959413 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.959456 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.959636 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.959703 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960232 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960249 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960313 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960405 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960529 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960587 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960692 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960651 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960864 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.960885 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961072 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961140 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961164 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961266 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961322 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961496 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961672 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961730 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.961768 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.961828 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.962352 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.964214 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.964877 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.964974 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.965208 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.965283 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.965490 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.965621 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.965918 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.965953 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.966019 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.966242 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.966799 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.966985 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.967301 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.968151 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.968533 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.968604 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.968812 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.968710 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.969004 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.969300 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.969423 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.969660 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.969749 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.970274 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.970609 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.970937 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.971037 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.971172 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.971322 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.971337 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.971405 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.972110 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.973145 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.973730 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.973946 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.974356 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.974707 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.975013 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.975033 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.975047 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.975333 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.975482 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.975912 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.976590 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.976834 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.976917 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.976941 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.976957 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.977428 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.978171 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.978512 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:56.478489388 +0000 UTC m=+20.967154020 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.978650 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:56.478629801 +0000 UTC m=+20.967294433 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.978757 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:56.478743834 +0000 UTC m=+20.967408466 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:55 crc kubenswrapper[4876]: E1205 08:15:55.979020 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:56.479005341 +0000 UTC m=+20.967669963 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.979142 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.971609 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.978514 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.980004 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.980484 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.981115 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.981216 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.981150 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.979766 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.978188 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.981306 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.981448 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.981480 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.982310 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.982172 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.982363 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.982566 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.983320 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.983620 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.983787 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.984089 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.984143 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.984240 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.984268 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.984704 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.984824 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.984842 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.984881 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.985080 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.985086 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.985977 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.986147 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.986752 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.986858 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.987260 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.987389 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.988342 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.988649 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.988601 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.989365 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.989612 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.989693 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.989892 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.989794 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.990135 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.990224 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.990264 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.990827 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.992411 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.992581 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.992595 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.992653 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.992664 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.993381 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.993693 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.994061 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.994169 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.994550 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.994511 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.995297 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.995532 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.995641 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.995808 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.996000 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.996542 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.996765 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.997150 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.997509 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:55 crc kubenswrapper[4876]: I1205 08:15:55.997752 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:55.999944 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.000180 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.001702 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.001996 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.002564 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.002888 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.012263 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.014254 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.024260 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.024947 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.028176 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.033701 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.041821 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.049646 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054157 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-etc-kubernetes\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054187 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/446e4833-20af-47c0-b4e0-7116c3972366-ovn-node-metrics-cert\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054210 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-config\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054226 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-cni-bin\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054240 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054253 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98p7w\" (UniqueName: \"kubernetes.io/projected/e7274e0d-7f30-495c-800a-5dcbf1656b0d-kube-api-access-98p7w\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054342 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-etc-kubernetes\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054417 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054420 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-cni-bin\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054494 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9djw\" (UniqueName: \"kubernetes.io/projected/77322cc8-c6ab-4250-8098-9938309f0af8-kube-api-access-v9djw\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054517 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-systemd-units\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054539 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-systemd\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054561 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-log-socket\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054581 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-env-overrides\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054602 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwhkw\" (UniqueName: \"kubernetes.io/projected/ae897c0d-b078-40d3-bbd4-aa2c279c89ec-kube-api-access-fwhkw\") pod \"node-resolver-n4dsp\" (UID: \"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\") " pod="openshift-dns/node-resolver-n4dsp" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054625 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/77322cc8-c6ab-4250-8098-9938309f0af8-mcd-auth-proxy-config\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054644 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-hostroot\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054667 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-etc-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054686 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ae897c0d-b078-40d3-bbd4-aa2c279c89ec-hosts-file\") pod \"node-resolver-n4dsp\" (UID: \"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\") " pod="openshift-dns/node-resolver-n4dsp" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054707 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-bin\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054731 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-conf-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054752 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-var-lib-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054775 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054798 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054821 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cni-binary-copy\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054842 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054877 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/77322cc8-c6ab-4250-8098-9938309f0af8-proxy-tls\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054918 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-kubelet\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054937 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-kubelet\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054959 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcg2f\" (UniqueName: \"kubernetes.io/projected/446e4833-20af-47c0-b4e0-7116c3972366-kube-api-access-fcg2f\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054989 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-system-cni-dir\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055008 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-cni-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055028 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-netns\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055034 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-config\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055070 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-ovn\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055047 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-ovn\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055151 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-script-lib\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055177 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cnibin\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.054774 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-hostroot\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055198 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e7274e0d-7f30-495c-800a-5dcbf1656b0d-cni-binary-copy\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055230 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-slash\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055270 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055289 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h79nh\" (UniqueName: \"kubernetes.io/projected/905a4c7a-1ee3-46ff-bd6b-04a34078015c-kube-api-access-h79nh\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055305 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-system-cni-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055320 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-cnibin\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055342 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-netd\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055365 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-netns\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055383 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-multus-certs\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055399 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-ovn-kubernetes\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055417 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-os-release\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055438 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-k8s-cni-cncf-io\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055457 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-socket-dir-parent\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055479 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/77322cc8-c6ab-4250-8098-9938309f0af8-rootfs\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055499 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055519 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-os-release\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055538 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-cni-multus\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055558 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-daemon-config\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055578 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-node-log\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055785 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055805 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055818 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055854 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055867 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055879 4876 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055890 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055917 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055928 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055939 4876 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055950 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055986 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.055997 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056009 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056085 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056096 4876 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056107 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056119 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056130 4876 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056142 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056154 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056166 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056179 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056190 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056197 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056201 4876 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056228 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-node-log\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056235 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056254 4876 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056267 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056278 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056289 4876 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056296 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ae897c0d-b078-40d3-bbd4-aa2c279c89ec-hosts-file\") pod \"node-resolver-n4dsp\" (UID: \"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\") " pod="openshift-dns/node-resolver-n4dsp" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056301 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056325 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056336 4876 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056347 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056358 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056369 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056380 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056391 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056407 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056417 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056431 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056442 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056453 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056464 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056474 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056486 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056498 4876 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056510 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056521 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056533 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056544 4876 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056555 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056567 4876 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056578 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056640 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056654 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056667 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056679 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056692 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056769 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e7274e0d-7f30-495c-800a-5dcbf1656b0d-cni-binary-copy\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056728 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057184 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057214 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057223 4876 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057233 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057242 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057253 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057262 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057272 4876 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057280 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057289 4876 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057297 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057306 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057318 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057327 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057329 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/77322cc8-c6ab-4250-8098-9938309f0af8-mcd-auth-proxy-config\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057337 4876 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057359 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-bin\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057371 4876 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057382 4876 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057386 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-conf-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057393 4876 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057404 4876 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057413 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057423 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-var-lib-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057442 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057448 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-systemd-units\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057468 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-slash\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057600 4876 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057656 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.057252 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-env-overrides\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058039 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-systemd\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058271 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-log-socket\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058323 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-system-cni-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058370 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-cnibin\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058397 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058408 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-k8s-cni-cncf-io\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058419 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/77322cc8-c6ab-4250-8098-9938309f0af8-rootfs\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058443 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-netd\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058451 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-socket-dir-parent\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058463 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-netns\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058484 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-run-multus-certs\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058506 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-ovn-kubernetes\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058541 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-os-release\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058558 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058591 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-system-cni-dir\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058641 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-kubelet\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058669 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-kubelet\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058843 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.058891 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-netns\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.059039 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-daemon-config\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.059069 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-multus-cni-dir\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.059103 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.056266 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-etc-openvswitch\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.059132 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-host-var-lib-cni-multus\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.059172 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e7274e0d-7f30-495c-800a-5dcbf1656b0d-os-release\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.059792 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.059920 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cni-binary-copy\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.059958 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/905a4c7a-1ee3-46ff-bd6b-04a34078015c-cnibin\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060108 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060127 4876 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060140 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060153 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060166 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060178 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060191 4876 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060203 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060216 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060228 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060242 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060255 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060267 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060289 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060302 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060315 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060327 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060343 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060355 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060366 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060379 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060390 4876 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060402 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060416 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060429 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060441 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060453 4876 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060465 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060477 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060489 4876 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060502 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060514 4876 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060526 4876 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060539 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060552 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061061 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/77322cc8-c6ab-4250-8098-9938309f0af8-proxy-tls\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060329 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-script-lib\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.060564 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061115 4876 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061127 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061139 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061167 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061179 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061195 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061209 4876 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061221 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061232 4876 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061244 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061259 4876 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061273 4876 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061285 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061297 4876 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061308 4876 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061321 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061333 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061345 4876 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061356 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061368 4876 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061380 4876 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061392 4876 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061404 4876 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061416 4876 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061428 4876 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061440 4876 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061452 4876 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061465 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061476 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061487 4876 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061500 4876 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061513 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061525 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061537 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061549 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061561 4876 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061573 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061587 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061599 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061633 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061647 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061659 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061672 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061684 4876 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061696 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061708 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061720 4876 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061732 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061744 4876 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061756 4876 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061768 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061780 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061792 4876 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061805 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061817 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061829 4876 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061851 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061864 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061876 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061889 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061921 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061933 4876 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061945 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.061959 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.071696 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98p7w\" (UniqueName: \"kubernetes.io/projected/e7274e0d-7f30-495c-800a-5dcbf1656b0d-kube-api-access-98p7w\") pod \"multus-jwrfx\" (UID: \"e7274e0d-7f30-495c-800a-5dcbf1656b0d\") " pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.075858 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.064386 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.091101 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.092069 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwhkw\" (UniqueName: \"kubernetes.io/projected/ae897c0d-b078-40d3-bbd4-aa2c279c89ec-kube-api-access-fwhkw\") pod \"node-resolver-n4dsp\" (UID: \"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\") " pod="openshift-dns/node-resolver-n4dsp" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.092109 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcg2f\" (UniqueName: \"kubernetes.io/projected/446e4833-20af-47c0-b4e0-7116c3972366-kube-api-access-fcg2f\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.093688 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/446e4833-20af-47c0-b4e0-7116c3972366-ovn-node-metrics-cert\") pod \"ovnkube-node-58npg\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.101415 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9djw\" (UniqueName: \"kubernetes.io/projected/77322cc8-c6ab-4250-8098-9938309f0af8-kube-api-access-v9djw\") pod \"machine-config-daemon-xj8qx\" (UID: \"77322cc8-c6ab-4250-8098-9938309f0af8\") " pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.130092 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jwrfx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.130124 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.130261 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.130565 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.131724 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.133214 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h79nh\" (UniqueName: \"kubernetes.io/projected/905a4c7a-1ee3-46ff-bd6b-04a34078015c-kube-api-access-h79nh\") pod \"multus-additional-cni-plugins-z44lm\" (UID: \"905a4c7a-1ee3-46ff-bd6b-04a34078015c\") " pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.140763 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-n4dsp" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.144242 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.152694 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.161429 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.170837 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: W1205 08:15:56.174738 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-bcd0bbd5e14ba1005ae1d6fd48749cfbc77dac9684f023dd8cd39cbe2ad62d0e WatchSource:0}: Error finding container bcd0bbd5e14ba1005ae1d6fd48749cfbc77dac9684f023dd8cd39cbe2ad62d0e: Status 404 returned error can't find the container with id bcd0bbd5e14ba1005ae1d6fd48749cfbc77dac9684f023dd8cd39cbe2ad62d0e Dec 05 08:15:56 crc kubenswrapper[4876]: W1205 08:15:56.181637 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae897c0d_b078_40d3_bbd4_aa2c279c89ec.slice/crio-b2a788ab9b6648c3e7dfc6cc6379f60aacbfcabeca9e6162c6df7b0345fac259 WatchSource:0}: Error finding container b2a788ab9b6648c3e7dfc6cc6379f60aacbfcabeca9e6162c6df7b0345fac259: Status 404 returned error can't find the container with id b2a788ab9b6648c3e7dfc6cc6379f60aacbfcabeca9e6162c6df7b0345fac259 Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.181663 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.189267 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.200373 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.212409 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.223626 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.239633 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.255639 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.268081 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.278310 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.285608 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.330918 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.340309 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.340702 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.359214 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.373560 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.387539 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.392602 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-z44lm" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.407184 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.425944 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.443803 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.453222 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.460867 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.464151 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.464354 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:15:57.464336262 +0000 UTC m=+21.953000884 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.493433 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.535295 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.565211 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.565259 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.565330 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565334 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.565354 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565408 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:57.565390197 +0000 UTC m=+22.054054819 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565466 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565497 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565514 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565560 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565571 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565520 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565617 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565526 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:57.56550888 +0000 UTC m=+22.054173572 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565690 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:57.565673564 +0000 UTC m=+22.054338196 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.565718 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:57.565705315 +0000 UTC m=+22.054370027 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.573830 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.615257 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.660827 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.695713 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.748300 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.778671 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.823003 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.823142 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: E1205 08:15:56.823446 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.855761 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.896061 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.937628 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.943537 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.943583 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.943596 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3f1fe7da8440b14709a0f2661782467fbaf97bb530b57233a7e27445ead20e90"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.945197 4876 generic.go:334] "Generic (PLEG): container finished" podID="905a4c7a-1ee3-46ff-bd6b-04a34078015c" containerID="0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5" exitCode=0 Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.945276 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" event={"ID":"905a4c7a-1ee3-46ff-bd6b-04a34078015c","Type":"ContainerDied","Data":"0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.945340 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" event={"ID":"905a4c7a-1ee3-46ff-bd6b-04a34078015c","Type":"ContainerStarted","Data":"4898ee5c10053e015a7936f4c92c43218919bae62323c8f2153947120ebe7644"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.946442 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-n4dsp" event={"ID":"ae897c0d-b078-40d3-bbd4-aa2c279c89ec","Type":"ContainerStarted","Data":"5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.946477 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-n4dsp" event={"ID":"ae897c0d-b078-40d3-bbd4-aa2c279c89ec","Type":"ContainerStarted","Data":"b2a788ab9b6648c3e7dfc6cc6379f60aacbfcabeca9e6162c6df7b0345fac259"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.947625 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.947650 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6b5b682e1656df281aa3ee8c0f1c6b935be7101995d4f3909edaee49e6d701bb"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.948609 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"bcd0bbd5e14ba1005ae1d6fd48749cfbc77dac9684f023dd8cd39cbe2ad62d0e"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.949473 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5" exitCode=0 Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.949527 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.949545 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"2a28c1f10fd27337083b72ff07e2ed349bceaf0d312551f0fa4d8a6d7c2fc807"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.952291 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.952318 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.952328 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"3d72e036cd37ba23280ada1389488fa3e0b57cc36a16536c7e0b28e46c3cb668"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.953633 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jwrfx" event={"ID":"e7274e0d-7f30-495c-800a-5dcbf1656b0d","Type":"ContainerStarted","Data":"3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.953658 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jwrfx" event={"ID":"e7274e0d-7f30-495c-800a-5dcbf1656b0d","Type":"ContainerStarted","Data":"a8e91019c103caa64adacfcfc88f8465f08b3abed1f1f83efaeff91f1ea571aa"} Dec 05 08:15:56 crc kubenswrapper[4876]: I1205 08:15:56.979130 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.013117 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.056954 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.100539 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.134161 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.137177 4876 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.138858 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.138886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.138917 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.139012 4876 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.197310 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.207126 4876 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.207406 4876 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.208384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.208424 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.208438 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.208455 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.208467 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.227453 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.230538 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.230569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.230582 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.230599 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.230611 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.241870 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.244882 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.244928 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.244940 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.244957 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.244969 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.257590 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.258319 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.261069 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.261105 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.261117 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.261133 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.261147 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.290888 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.303256 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.303495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.303822 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.304587 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.304722 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.305211 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.326201 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.326312 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.327660 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.327683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.327691 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.327704 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.327713 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.339610 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.375067 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.416896 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.430062 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.430109 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.430121 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.430148 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.430160 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.454249 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.475689 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.475846 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:15:59.475823095 +0000 UTC m=+23.964487717 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.496428 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.520476 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tlgq9"] Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.520914 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.532840 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.532881 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.532891 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.532924 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.532934 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.536470 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.548137 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.566933 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.577045 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.577085 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.577104 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-host\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.577124 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnq84\" (UniqueName: \"kubernetes.io/projected/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-kube-api-access-hnq84\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.577144 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577156 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577184 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577195 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577235 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:59.577221449 +0000 UTC m=+24.065886071 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577237 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577263 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:59.57725732 +0000 UTC m=+24.065921942 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.577161 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-serviceca\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577268 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.577293 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577305 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:59.577292721 +0000 UTC m=+24.065957333 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577367 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577377 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577383 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.577402 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 08:15:59.577396364 +0000 UTC m=+24.066060986 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.586341 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.606574 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.635102 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.635153 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.635171 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.635193 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.635207 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.654846 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.677786 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-host\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.677834 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnq84\" (UniqueName: \"kubernetes.io/projected/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-kube-api-access-hnq84\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.677868 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-serviceca\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.677866 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-host\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.678968 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-serviceca\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.701364 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.729544 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnq84\" (UniqueName: \"kubernetes.io/projected/e9911fb0-ecb1-429a-bf57-7c2ea8d2c505-kube-api-access-hnq84\") pod \"node-ca-tlgq9\" (UID: \"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\") " pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.737584 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.737647 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.737666 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.737695 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.737717 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.757524 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.816717 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.823022 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.823185 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.823024 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:57 crc kubenswrapper[4876]: E1205 08:15:57.823307 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.827090 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.828134 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.829233 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.829837 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.830779 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.831298 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.832050 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.832779 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.833832 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.834383 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.834893 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.835893 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.836580 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.837086 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.837960 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.838261 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.839303 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.839485 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.839559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.839585 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.839616 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.839637 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.840528 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.841066 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.841758 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.843081 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.843690 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.844371 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.845258 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.846047 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.847200 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.847824 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.848867 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.849344 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.850429 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.850928 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.851369 4876 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.851813 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.853421 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.853972 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.854558 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tlgq9" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.854784 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.856603 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.857562 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.858328 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.859209 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.860127 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.860827 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.861696 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.862549 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.865398 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.866531 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.868728 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.870737 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.872349 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.873472 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: W1205 08:15:57.873868 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9911fb0_ecb1_429a_bf57_7c2ea8d2c505.slice/crio-435acf6b207713c82458e86b6cea0c059c79ebe339009c96d03ea0077b84597e WatchSource:0}: Error finding container 435acf6b207713c82458e86b6cea0c059c79ebe339009c96d03ea0077b84597e: Status 404 returned error can't find the container with id 435acf6b207713c82458e86b6cea0c059c79ebe339009c96d03ea0077b84597e Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.875609 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.876792 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.878221 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.878600 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.879785 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.880814 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.916174 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.942627 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.942677 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.942693 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.942715 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.942737 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:57Z","lastTransitionTime":"2025-12-05T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.958046 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.960411 4876 generic.go:334] "Generic (PLEG): container finished" podID="905a4c7a-1ee3-46ff-bd6b-04a34078015c" containerID="dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9" exitCode=0 Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.960465 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" event={"ID":"905a4c7a-1ee3-46ff-bd6b-04a34078015c","Type":"ContainerDied","Data":"dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.962135 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tlgq9" event={"ID":"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505","Type":"ContainerStarted","Data":"435acf6b207713c82458e86b6cea0c059c79ebe339009c96d03ea0077b84597e"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.969852 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.969884 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.969937 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.969949 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.969958 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.969967 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d"} Dec 05 08:15:57 crc kubenswrapper[4876]: I1205 08:15:57.997358 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:57Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.041822 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.045501 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.045544 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.045555 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.045571 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.045583 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.081544 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.117763 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.147873 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.147914 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.147927 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.147940 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.147950 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.162917 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.196723 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.237980 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.250272 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.250317 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.250328 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.250346 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.250360 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.277301 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.323987 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.353015 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.353059 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.353071 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.353087 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.353099 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.359028 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.403331 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.442458 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.457095 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.457159 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.457182 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.457206 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.457406 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.486038 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.523179 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.559649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.559889 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.560009 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.560140 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.560429 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.560428 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.599868 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.638849 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.663590 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.663652 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.663674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.663705 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.663726 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.686768 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.719837 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.762765 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.767404 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.767473 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.767491 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.767518 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.767537 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.798625 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.823674 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:15:58 crc kubenswrapper[4876]: E1205 08:15:58.823857 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.842246 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.869551 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.869629 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.869644 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.869663 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.869678 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.881361 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.922169 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.972273 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.972321 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.972337 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.972356 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.972372 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:58Z","lastTransitionTime":"2025-12-05T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.975328 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tlgq9" event={"ID":"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505","Type":"ContainerStarted","Data":"5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.979242 4876 generic.go:334] "Generic (PLEG): container finished" podID="905a4c7a-1ee3-46ff-bd6b-04a34078015c" containerID="cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943" exitCode=0 Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.979287 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" event={"ID":"905a4c7a-1ee3-46ff-bd6b-04a34078015c","Type":"ContainerDied","Data":"cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943"} Dec 05 08:15:58 crc kubenswrapper[4876]: I1205 08:15:58.996155 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.012106 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.039536 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.074687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.074713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.074722 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.074735 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.074743 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.077319 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.128716 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.158652 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.177437 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.177481 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.177493 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.177537 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.177549 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.198175 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.239959 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.278398 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.280108 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.280176 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.280198 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.280226 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.280249 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.318545 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.354647 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.382015 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.382057 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.382069 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.382085 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.382096 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.401587 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.440950 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.477526 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.484595 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.484656 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.484672 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.484693 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.484710 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.498008 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.498188 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:16:03.498159587 +0000 UTC m=+27.986824249 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.519530 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.563414 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.587814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.587886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.587933 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.587957 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.587974 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.596023 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.599385 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.599439 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.599500 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599560 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599650 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:03.599628393 +0000 UTC m=+28.088293015 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.599536 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599678 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599741 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599779 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599783 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:03.599761016 +0000 UTC m=+28.088425638 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599795 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599809 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599850 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:03.599828778 +0000 UTC m=+28.088493460 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599855 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.599881 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.600032 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:03.599989812 +0000 UTC m=+28.088654474 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.634802 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.674675 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.691227 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.691278 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.691299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.691322 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.691339 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.725435 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.767328 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.793299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.793329 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.793338 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.793351 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.793362 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.799171 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.823575 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.823572 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.823712 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:15:59 crc kubenswrapper[4876]: E1205 08:15:59.823800 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.840451 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.873724 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.895542 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.895576 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.895587 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.895602 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.895610 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.916260 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.957279 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.982616 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.985475 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.987570 4876 generic.go:334] "Generic (PLEG): container finished" podID="905a4c7a-1ee3-46ff-bd6b-04a34078015c" containerID="1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7" exitCode=0 Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.987627 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" event={"ID":"905a4c7a-1ee3-46ff-bd6b-04a34078015c","Type":"ContainerDied","Data":"1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7"} Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.996233 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:15:59Z is after 2025-08-24T17:21:41Z" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.997024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.997051 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.997060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.997074 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:15:59 crc kubenswrapper[4876]: I1205 08:15:59.997083 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:15:59Z","lastTransitionTime":"2025-12-05T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.036658 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.075208 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.099419 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.099511 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.099552 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.099571 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.099592 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.113228 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.152262 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.199994 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.202128 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.202166 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.202175 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.202191 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.202201 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.237241 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.278458 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.304267 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.304330 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.304349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.304372 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.304392 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.321804 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.364107 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.402858 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.406577 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.406767 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.406954 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.407100 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.407288 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.440034 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.490876 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.511180 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.511358 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.511439 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.511517 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.511589 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.525608 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.559774 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.598598 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:00Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.614056 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.614113 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.614122 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.614135 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.614144 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.716814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.716892 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.716934 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.716952 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.717009 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.820296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.820359 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.820373 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.820394 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.820429 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.823476 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:00 crc kubenswrapper[4876]: E1205 08:16:00.823654 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.923978 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.924032 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.924046 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.924069 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.924086 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:00Z","lastTransitionTime":"2025-12-05T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.995583 4876 generic.go:334] "Generic (PLEG): container finished" podID="905a4c7a-1ee3-46ff-bd6b-04a34078015c" containerID="2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c" exitCode=0 Dec 05 08:16:00 crc kubenswrapper[4876]: I1205 08:16:00.996099 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" event={"ID":"905a4c7a-1ee3-46ff-bd6b-04a34078015c","Type":"ContainerDied","Data":"2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.018310 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.027683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.027720 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.027729 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.027745 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.027756 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.035997 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.049268 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.064280 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.077124 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.097077 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.114940 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.129283 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.129858 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.129884 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.129911 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.129928 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.129937 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.142650 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.156925 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.169694 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.180763 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.190812 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.200352 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:01Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.233043 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.233083 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.233093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.233109 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.233122 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.335547 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.335598 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.335615 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.335639 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.335656 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.438559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.439069 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.439089 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.439117 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.439136 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.543101 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.543179 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.543203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.543232 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.543253 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.645622 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.645676 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.645687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.645703 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.645716 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.747946 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.747996 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.748011 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.748031 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.748046 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.823450 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.823468 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:01 crc kubenswrapper[4876]: E1205 08:16:01.823633 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:01 crc kubenswrapper[4876]: E1205 08:16:01.823828 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.850888 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.850964 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.850976 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.850994 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.851010 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.952877 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.952942 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.952954 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.952971 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:01 crc kubenswrapper[4876]: I1205 08:16:01.952982 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:01Z","lastTransitionTime":"2025-12-05T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.006422 4876 generic.go:334] "Generic (PLEG): container finished" podID="905a4c7a-1ee3-46ff-bd6b-04a34078015c" containerID="089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea" exitCode=0 Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.006482 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" event={"ID":"905a4c7a-1ee3-46ff-bd6b-04a34078015c","Type":"ContainerDied","Data":"089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.024339 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.049155 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.055361 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.055404 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.055416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.055433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.055446 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.068248 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.090018 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.105547 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.118955 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.136593 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.150714 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.159159 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.159211 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.159230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.159255 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.159337 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.167067 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.184953 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.196434 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.206590 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.216617 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.233571 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:02Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.261187 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.261221 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.261232 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.261248 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.261258 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.364444 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.364493 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.364513 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.364537 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.364554 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.467253 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.467325 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.467351 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.467381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.467404 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.570814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.570852 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.570864 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.570879 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.570890 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.673479 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.673551 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.673573 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.673597 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.673614 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.776929 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.776984 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.776996 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.777016 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.777034 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.822820 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:02 crc kubenswrapper[4876]: E1205 08:16:02.822960 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.879209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.879266 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.879280 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.879296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.879308 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.982072 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.982131 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.982143 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.982161 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:02 crc kubenswrapper[4876]: I1205 08:16:02.982175 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:02Z","lastTransitionTime":"2025-12-05T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.014487 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.016502 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.016544 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.018262 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" event={"ID":"905a4c7a-1ee3-46ff-bd6b-04a34078015c","Type":"ContainerStarted","Data":"d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.034811 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.045281 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.047226 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.049229 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.067598 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.085067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.085105 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.085113 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.085127 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.085136 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.092567 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.104954 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.118105 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.134310 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.155287 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.176426 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.187367 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.187446 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.187472 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.187502 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.187525 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.189273 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.205204 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.219570 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.232720 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.249266 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.266316 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.284523 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.289771 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.289829 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.289841 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.289863 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.289875 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.300181 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.317114 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.333765 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.365721 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.388463 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.393112 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.393212 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.393234 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.393266 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.393290 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.408885 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.426722 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.445352 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.461519 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.477840 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.491005 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.495472 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.495517 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.495533 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.495554 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.495571 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.505799 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.536342 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.536444 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:16:11.536423885 +0000 UTC m=+36.025088527 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.597958 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.598007 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.598020 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.598037 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.598049 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.636998 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.637052 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.637091 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.637118 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637186 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637229 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637281 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:11.637257274 +0000 UTC m=+36.125921906 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637302 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:11.637293065 +0000 UTC m=+36.125957697 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637323 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637338 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637350 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637387 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:11.637371047 +0000 UTC m=+36.126035769 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637489 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637510 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637523 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.637562 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:11.637549172 +0000 UTC m=+36.126213814 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.700165 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.700238 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.700256 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.700282 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.700300 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.803305 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.803384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.803410 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.803441 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.803463 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.822955 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.823001 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.823167 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:03 crc kubenswrapper[4876]: E1205 08:16:03.823309 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.906817 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.906866 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.906884 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.906934 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.906952 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:03Z","lastTransitionTime":"2025-12-05T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:03 crc kubenswrapper[4876]: I1205 08:16:03.914670 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.010349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.010398 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.010416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.010438 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.010454 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.113169 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.113249 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.113275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.113307 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.113332 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.215629 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.215695 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.215716 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.215742 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.215759 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.318888 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.318943 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.318955 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.318974 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.318984 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.421025 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.421065 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.421076 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.421093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.421107 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.524103 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.524354 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.524376 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.524408 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.524431 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.626566 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.626603 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.626613 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.626630 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.626643 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.729583 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.729638 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.729655 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.729680 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.729710 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.823994 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:04 crc kubenswrapper[4876]: E1205 08:16:04.824195 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.832399 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.832796 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.832890 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.833025 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.833143 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.937182 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.937256 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.937278 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.937308 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:04 crc kubenswrapper[4876]: I1205 08:16:04.937332 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:04Z","lastTransitionTime":"2025-12-05T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.027205 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/0.log" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.031447 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba" exitCode=1 Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.031502 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.032453 4876 scope.go:117] "RemoveContainer" containerID="9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.040526 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.040581 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.040607 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.040633 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.040652 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.050532 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.108265 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.135982 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:04Z\\\",\\\"message\\\":\\\" 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 08:16:04.741309 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 08:16:04.741314 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:04.741216 6172 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:04.741353 6172 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:04.741401 6172 factory.go:656] Stopping watch factory\\\\nI1205 08:16:04.741426 6172 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:04.741438 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:04.741270 6172 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:04.741465 6172 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 08:16:04.741597 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:04.741616 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:04.741631 6172 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.142673 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.142708 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.142718 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.142736 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.142746 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.157402 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.171506 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.187084 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.205523 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.219980 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.234882 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.245022 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.245067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.245079 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.245093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.245103 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.251532 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.266244 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.279713 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.292870 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.306329 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.347415 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.347450 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.347460 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.347476 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.347485 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.450232 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.450273 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.450284 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.450301 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.450312 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.553429 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.553487 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.553504 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.553525 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.553541 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.655485 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.655529 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.655542 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.655559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.655570 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.758310 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.758342 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.758349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.758362 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.758369 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.823742 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.823770 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:05 crc kubenswrapper[4876]: E1205 08:16:05.824446 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:05 crc kubenswrapper[4876]: E1205 08:16:05.824641 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.844494 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.860836 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.860884 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.860915 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.860933 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.860946 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.861406 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.878392 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.902063 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.915207 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.943502 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:04Z\\\",\\\"message\\\":\\\" 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 08:16:04.741309 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 08:16:04.741314 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:04.741216 6172 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:04.741353 6172 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:04.741401 6172 factory.go:656] Stopping watch factory\\\\nI1205 08:16:04.741426 6172 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:04.741438 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:04.741270 6172 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:04.741465 6172 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 08:16:04.741597 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:04.741616 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:04.741631 6172 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.962579 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.965330 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.965400 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.965425 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.965453 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.965470 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:05Z","lastTransitionTime":"2025-12-05T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:05 crc kubenswrapper[4876]: I1205 08:16:05.982809 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:05.997093 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.016204 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.029741 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.037174 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/0.log" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.040359 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.040950 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.043785 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.054487 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.066864 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.068338 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.068418 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.068433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.068460 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.068492 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.084958 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.098891 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.111649 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.122389 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.141188 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.155217 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.171242 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.171289 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.171302 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.171324 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.171343 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.172204 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.181535 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.194736 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.207693 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.220118 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.236331 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.252790 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.271503 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:04Z\\\",\\\"message\\\":\\\" 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 08:16:04.741309 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 08:16:04.741314 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:04.741216 6172 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:04.741353 6172 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:04.741401 6172 factory.go:656] Stopping watch factory\\\\nI1205 08:16:04.741426 6172 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:04.741438 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:04.741270 6172 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:04.741465 6172 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 08:16:04.741597 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:04.741616 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:04.741631 6172 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.273660 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.273706 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.273716 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.273732 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.273743 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.376528 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.376592 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.376605 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.376625 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.376637 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.479361 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.479429 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.479448 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.479475 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.479494 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.582709 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.582757 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.582769 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.582789 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.582807 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.686400 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.686463 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.686482 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.686509 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.686539 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.792729 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.792793 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.792816 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.792847 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.792876 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.823685 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:06 crc kubenswrapper[4876]: E1205 08:16:06.823952 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.896496 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.896553 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.896579 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.896607 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:06 crc kubenswrapper[4876]: I1205 08:16:06.896629 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:06Z","lastTransitionTime":"2025-12-05T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.000263 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.000309 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.000319 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.000340 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.000351 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.046456 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/1.log" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.047567 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/0.log" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.052182 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50" exitCode=1 Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.052222 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.052269 4876 scope.go:117] "RemoveContainer" containerID="9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.058460 4876 scope.go:117] "RemoveContainer" containerID="1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50" Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.058748 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\"" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.072924 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.095756 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.109612 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.109663 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.109681 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.109706 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.109724 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.128536 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.143693 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.156656 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.176407 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.195716 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.208925 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.212209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.212237 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.212245 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.212258 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.212266 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.223578 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.243568 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:04Z\\\",\\\"message\\\":\\\" 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 08:16:04.741309 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 08:16:04.741314 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:04.741216 6172 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:04.741353 6172 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:04.741401 6172 factory.go:656] Stopping watch factory\\\\nI1205 08:16:04.741426 6172 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:04.741438 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:04.741270 6172 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:04.741465 6172 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 08:16:04.741597 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:04.741616 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:04.741631 6172 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.261488 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.276307 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.289042 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.311930 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.314196 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.314232 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.314246 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.314262 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.314275 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.338929 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.338965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.338975 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.338988 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.338996 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.352746 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.355541 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.355569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.355580 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.355596 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.355607 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.365965 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.369457 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.369492 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.369502 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.369516 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.369525 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.380310 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.382864 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.382916 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.382928 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.382943 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.382953 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.395477 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.399176 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.399233 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.399250 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.399274 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.399290 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.413250 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.413409 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.416689 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.416739 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.416750 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.416765 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.416776 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.519821 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.519887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.519940 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.519965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.519981 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.623429 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.623500 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.623523 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.623551 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.623573 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.726399 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd"] Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.727240 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.727548 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.727603 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.727627 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.727654 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.727681 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.730293 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.730894 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.751872 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.775836 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.780388 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.780471 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.780512 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwvp\" (UniqueName: \"kubernetes.io/projected/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-kube-api-access-djwvp\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.780633 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.801675 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.823222 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.823251 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.823477 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:07 crc kubenswrapper[4876]: E1205 08:16:07.823617 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.830347 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.830402 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.830427 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.830451 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.830467 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.831359 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.858300 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.873389 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.881371 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.881702 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.881761 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.881801 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwvp\" (UniqueName: \"kubernetes.io/projected/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-kube-api-access-djwvp\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.881830 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.883008 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.883032 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.890160 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.899082 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.903337 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.905106 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwvp\" (UniqueName: \"kubernetes.io/projected/1229eb9b-4d2a-4301-bbdb-9b37e590ac11-kube-api-access-djwvp\") pod \"ovnkube-control-plane-749d76644c-kl7xd\" (UID: \"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.915665 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.925095 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.933031 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.933085 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.933096 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.933120 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.933132 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:07Z","lastTransitionTime":"2025-12-05T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.934917 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.944873 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.961302 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9777f338e67e772d14997505ff60699972d8affa566de1c8941e34c448141cba\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:04Z\\\",\\\"message\\\":\\\" 6172 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 08:16:04.741309 6172 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 08:16:04.741314 6172 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:04.741216 6172 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:04.741353 6172 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:04.741401 6172 factory.go:656] Stopping watch factory\\\\nI1205 08:16:04.741426 6172 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:04.741438 6172 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:04.741270 6172 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:04.741465 6172 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 08:16:04.741597 6172 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:04.741616 6172 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:04.741631 6172 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:07 crc kubenswrapper[4876]: I1205 08:16:07.976535 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:07Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.035462 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.035486 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.035495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.035528 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.035537 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.048851 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.056472 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/1.log" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.059488 4876 scope.go:117] "RemoveContainer" containerID="1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50" Dec 05 08:16:08 crc kubenswrapper[4876]: E1205 08:16:08.059671 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\"" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" Dec 05 08:16:08 crc kubenswrapper[4876]: W1205 08:16:08.071649 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1229eb9b_4d2a_4301_bbdb_9b37e590ac11.slice/crio-a5f1f0a8cdba1eb1e70e41a3d75c3f263f75f78204e84f69c0262c06cf1d58b0 WatchSource:0}: Error finding container a5f1f0a8cdba1eb1e70e41a3d75c3f263f75f78204e84f69c0262c06cf1d58b0: Status 404 returned error can't find the container with id a5f1f0a8cdba1eb1e70e41a3d75c3f263f75f78204e84f69c0262c06cf1d58b0 Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.081397 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.103246 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.105061 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.123356 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.136546 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.138669 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.138742 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.138759 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.138777 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.138789 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.147955 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.162512 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.176409 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.187304 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.197493 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.214074 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.241394 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.241433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.241445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.241460 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.241472 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.244709 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.256817 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.269388 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.278551 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.290891 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.306736 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.317354 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.328587 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.341011 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.344953 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.345030 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.345050 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.345078 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.345097 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.356547 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.376972 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.388194 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.400697 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.413094 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.429128 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.441114 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.446714 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.446741 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.446750 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.446777 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.446787 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.454371 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.465420 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.480527 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-7wg59"] Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.481178 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:08 crc kubenswrapper[4876]: E1205 08:16:08.481262 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.481372 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.493459 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.511823 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.524344 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.544084 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.548630 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.548680 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.548692 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.548708 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.548719 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.560421 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.578192 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.590165 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.592084 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.592120 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqmth\" (UniqueName: \"kubernetes.io/projected/79225fe3-5793-4316-92da-c227edb92d53-kube-api-access-gqmth\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.599465 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.612425 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.625825 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.641295 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.651746 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.651802 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.651821 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.651846 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.651863 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.660419 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.670487 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.684176 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.693538 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.693590 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqmth\" (UniqueName: \"kubernetes.io/projected/79225fe3-5793-4316-92da-c227edb92d53-kube-api-access-gqmth\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:08 crc kubenswrapper[4876]: E1205 08:16:08.693776 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:08 crc kubenswrapper[4876]: E1205 08:16:08.693865 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs podName:79225fe3-5793-4316-92da-c227edb92d53 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:09.193829571 +0000 UTC m=+33.682494233 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs") pod "network-metrics-daemon-7wg59" (UID: "79225fe3-5793-4316-92da-c227edb92d53") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.699056 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.720649 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.725018 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqmth\" (UniqueName: \"kubernetes.io/projected/79225fe3-5793-4316-92da-c227edb92d53-kube-api-access-gqmth\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.736804 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:08Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.755070 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.755142 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.755164 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.755260 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.755290 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.823767 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:08 crc kubenswrapper[4876]: E1205 08:16:08.823963 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.857762 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.857809 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.857821 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.857840 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.857858 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.961758 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.961807 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.961823 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.961843 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:08 crc kubenswrapper[4876]: I1205 08:16:08.961857 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:08Z","lastTransitionTime":"2025-12-05T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.064074 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.064104 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.064113 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.064127 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.064135 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.064651 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" event={"ID":"1229eb9b-4d2a-4301-bbdb-9b37e590ac11","Type":"ContainerStarted","Data":"6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.064676 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" event={"ID":"1229eb9b-4d2a-4301-bbdb-9b37e590ac11","Type":"ContainerStarted","Data":"68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.064685 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" event={"ID":"1229eb9b-4d2a-4301-bbdb-9b37e590ac11","Type":"ContainerStarted","Data":"a5f1f0a8cdba1eb1e70e41a3d75c3f263f75f78204e84f69c0262c06cf1d58b0"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.086386 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.099774 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.110786 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.120081 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.128539 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.138483 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.152754 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.166565 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.166616 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.166627 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.166644 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.166656 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.168935 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.182011 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.191959 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.198767 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:09 crc kubenswrapper[4876]: E1205 08:16:09.198872 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:09 crc kubenswrapper[4876]: E1205 08:16:09.198932 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs podName:79225fe3-5793-4316-92da-c227edb92d53 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:10.198918891 +0000 UTC m=+34.687583513 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs") pod "network-metrics-daemon-7wg59" (UID: "79225fe3-5793-4316-92da-c227edb92d53") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.204655 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.223524 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.236158 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.249712 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.259660 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.269184 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.269253 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.269272 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.269297 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.269318 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.277516 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.372887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.372959 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.372974 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.372992 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.373004 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.475713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.476026 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.476036 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.476049 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.476063 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.578565 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.578651 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.578680 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.578713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.578737 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.681620 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.681692 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.681710 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.682172 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.682221 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.785119 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.785252 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.785277 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.785305 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.785329 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.823971 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.824011 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:09 crc kubenswrapper[4876]: E1205 08:16:09.824194 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.824354 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:09 crc kubenswrapper[4876]: E1205 08:16:09.824553 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:09 crc kubenswrapper[4876]: E1205 08:16:09.824716 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.887672 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.887741 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.887763 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.887791 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.887812 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.990749 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.990818 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.990837 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.990861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:09 crc kubenswrapper[4876]: I1205 08:16:09.990881 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:09Z","lastTransitionTime":"2025-12-05T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.093787 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.093845 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.093862 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.093884 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.093942 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.196551 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.196612 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.196633 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.196656 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.196672 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.208895 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:10 crc kubenswrapper[4876]: E1205 08:16:10.209002 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:10 crc kubenswrapper[4876]: E1205 08:16:10.209075 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs podName:79225fe3-5793-4316-92da-c227edb92d53 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:12.209059881 +0000 UTC m=+36.697724503 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs") pod "network-metrics-daemon-7wg59" (UID: "79225fe3-5793-4316-92da-c227edb92d53") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.298870 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.299003 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.299020 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.299046 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.299065 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.401446 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.401498 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.401515 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.401538 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.401555 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.504626 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.504694 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.504716 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.504745 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.504770 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.607582 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.607646 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.607664 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.607687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.607704 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.711318 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.711362 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.711377 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.711399 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.711416 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.814171 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.814205 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.814217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.814253 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.814264 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.823026 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:10 crc kubenswrapper[4876]: E1205 08:16:10.823189 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.917197 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.917253 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.917275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.917303 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:10 crc kubenswrapper[4876]: I1205 08:16:10.917325 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:10Z","lastTransitionTime":"2025-12-05T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.020355 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.020416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.020434 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.020457 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.020474 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.123092 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.123151 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.123167 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.123189 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.123207 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.226350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.226427 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.226461 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.226484 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.226501 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.328794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.328870 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.328893 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.328965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.328989 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.431838 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.431942 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.431968 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.431998 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.432021 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.534726 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.534755 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.534764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.534778 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.534787 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.621363 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.621546 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:16:27.621515271 +0000 UTC m=+52.110179933 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.637422 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.637498 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.637545 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.637583 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.637667 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.723599 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.723696 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.723743 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.723782 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.723782 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.723944 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:27.723873899 +0000 UTC m=+52.212538561 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.723995 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724023 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724032 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724053 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724074 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724081 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724097 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724099 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:27.724074904 +0000 UTC m=+52.212739566 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724170 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:27.724151466 +0000 UTC m=+52.212816118 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.724193 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:27.724181357 +0000 UTC m=+52.212846009 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.740796 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.741000 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.741041 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.741075 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.741099 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.823241 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.823391 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.823439 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.823473 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.824061 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:11 crc kubenswrapper[4876]: E1205 08:16:11.824278 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.844023 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.844079 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.844098 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.844123 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.844142 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.946826 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.946952 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.946977 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.947010 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:11 crc kubenswrapper[4876]: I1205 08:16:11.947033 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:11Z","lastTransitionTime":"2025-12-05T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.049586 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.049658 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.049666 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.049680 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.049688 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.152794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.152872 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.152893 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.152971 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.152995 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.230472 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:12 crc kubenswrapper[4876]: E1205 08:16:12.230651 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:12 crc kubenswrapper[4876]: E1205 08:16:12.230742 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs podName:79225fe3-5793-4316-92da-c227edb92d53 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:16.230714815 +0000 UTC m=+40.719379467 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs") pod "network-metrics-daemon-7wg59" (UID: "79225fe3-5793-4316-92da-c227edb92d53") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.255792 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.255844 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.255865 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.255893 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.255950 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.359496 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.359978 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.360115 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.360226 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.360330 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.462741 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.462843 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.462861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.462892 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.462931 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.566227 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.566276 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.566298 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.566325 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.566349 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.669313 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.669417 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.669443 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.669474 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.669496 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.771833 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.771974 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.772004 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.772035 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.772054 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.823404 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:12 crc kubenswrapper[4876]: E1205 08:16:12.823574 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.875176 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.875234 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.875253 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.875278 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.875295 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.978751 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.978823 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.978846 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.978876 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:12 crc kubenswrapper[4876]: I1205 08:16:12.978960 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:12Z","lastTransitionTime":"2025-12-05T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.081789 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.081852 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.081872 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.081895 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.081983 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.185495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.185550 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.185569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.185615 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.185639 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.289156 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.289234 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.289258 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.289291 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.289312 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.391777 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.391861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.391886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.391948 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.391968 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.494431 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.494485 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.494502 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.494526 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.494544 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.597693 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.597764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.597782 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.597809 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.597827 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.700659 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.700754 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.700772 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.700796 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.700814 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.804228 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.804303 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.804327 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.804354 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.804374 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.823004 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.823058 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.823099 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:13 crc kubenswrapper[4876]: E1205 08:16:13.823206 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:13 crc kubenswrapper[4876]: E1205 08:16:13.823464 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:13 crc kubenswrapper[4876]: E1205 08:16:13.823708 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.907621 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.907691 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.907718 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.907746 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:13 crc kubenswrapper[4876]: I1205 08:16:13.907770 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:13Z","lastTransitionTime":"2025-12-05T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.010691 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.010780 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.010798 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.010848 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.010873 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.114566 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.114744 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.115042 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.115081 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.115256 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.218006 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.218052 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.218068 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.218126 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.218146 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.321612 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.321654 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.321671 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.321693 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.321710 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.424604 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.424729 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.424756 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.424783 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.424804 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.528196 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.528276 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.528335 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.528360 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.528953 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.632485 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.632550 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.632571 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.632598 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.632622 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.736123 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.736482 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.736976 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.737237 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.737521 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.823142 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:14 crc kubenswrapper[4876]: E1205 08:16:14.823349 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.840089 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.840233 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.840260 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.840292 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.840343 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.943618 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.943689 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.943710 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.943735 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:14 crc kubenswrapper[4876]: I1205 08:16:14.943752 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:14Z","lastTransitionTime":"2025-12-05T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.047218 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.047281 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.047299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.047323 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.047341 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.149961 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.150024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.150045 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.150073 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.150099 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.253337 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.253411 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.253435 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.253461 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.253481 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.356992 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.357065 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.357085 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.357110 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.357130 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.459647 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.459689 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.459700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.459717 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.459736 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.563181 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.563230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.563248 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.563272 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.563290 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.665731 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.665790 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.665807 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.665832 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.665851 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.768685 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.768753 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.768771 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.768794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.768812 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.823499 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.823666 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:15 crc kubenswrapper[4876]: E1205 08:16:15.823794 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.823886 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:15 crc kubenswrapper[4876]: E1205 08:16:15.824005 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:15 crc kubenswrapper[4876]: E1205 08:16:15.824093 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.841542 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:15Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.858205 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:15Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.871572 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.871643 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.871663 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.871691 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.871711 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.876956 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:15Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.896530 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:15Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.920532 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:15Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.940860 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:15Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.962444 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:15Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.974136 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.974201 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.974226 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.974251 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.974268 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:15Z","lastTransitionTime":"2025-12-05T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:15 crc kubenswrapper[4876]: I1205 08:16:15.987086 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:15Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.004373 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:16Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.022462 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:16Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.044536 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:16Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.062303 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:16Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.077237 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.077293 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.077308 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.077331 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.077350 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.083675 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:16Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.104983 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:16Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.118636 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:16Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.130320 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:16Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.179415 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.179732 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.180058 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.180196 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.180288 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.276665 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:16 crc kubenswrapper[4876]: E1205 08:16:16.276940 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:16 crc kubenswrapper[4876]: E1205 08:16:16.277316 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs podName:79225fe3-5793-4316-92da-c227edb92d53 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:24.277294249 +0000 UTC m=+48.765958881 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs") pod "network-metrics-daemon-7wg59" (UID: "79225fe3-5793-4316-92da-c227edb92d53") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.283545 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.283601 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.283617 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.283665 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.283683 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.386666 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.386729 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.386746 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.386771 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.386789 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.489442 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.489508 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.489529 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.489557 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.489576 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.592567 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.592636 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.592653 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.592677 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.592695 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.696312 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.696371 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.696424 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.696452 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.696502 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.799214 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.799291 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.799315 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.799345 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.799371 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.823045 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:16 crc kubenswrapper[4876]: E1205 08:16:16.823251 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.902836 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.902941 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.902967 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.902997 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:16 crc kubenswrapper[4876]: I1205 08:16:16.903020 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:16Z","lastTransitionTime":"2025-12-05T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.005975 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.006064 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.006081 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.006104 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.006122 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.108887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.108985 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.109007 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.109034 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.109057 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.211607 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.211657 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.211670 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.211687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.211697 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.314868 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.314971 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.314998 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.315027 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.315049 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.418177 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.418241 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.418260 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.418286 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.418305 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.522853 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.523276 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.523478 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.523630 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.523778 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.569125 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.569180 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.569197 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.569221 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.569237 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.591118 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:17Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.596559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.596627 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.596650 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.596677 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.596695 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.616288 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:17Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.621314 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.621360 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.621375 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.621396 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.621410 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.640034 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:17Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.645108 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.645330 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.645487 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.645639 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.645787 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.666096 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:17Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.672206 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.672448 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.672641 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.672858 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.673090 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.696355 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:17Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.696697 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.699063 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.699302 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.699495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.699651 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.699794 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.807019 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.807089 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.807107 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.807131 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.807149 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.823721 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.823819 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.823860 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.823952 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.823996 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:17 crc kubenswrapper[4876]: E1205 08:16:17.824105 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.910238 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.910290 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.910303 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.910322 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:17 crc kubenswrapper[4876]: I1205 08:16:17.910335 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:17Z","lastTransitionTime":"2025-12-05T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.013832 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.013881 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.013938 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.013962 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.013978 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.116564 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.116814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.116999 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.117207 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.117409 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.220968 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.221242 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.221488 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.221708 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.221933 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.325108 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.325177 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.325203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.325227 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.325244 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.428865 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.428996 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.429024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.429049 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.429067 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.533377 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.533466 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.533490 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.533521 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.533545 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.636476 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.636540 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.636562 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.636590 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.636612 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.740203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.740259 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.740277 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.740306 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.740323 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.823665 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:18 crc kubenswrapper[4876]: E1205 08:16:18.824048 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.825241 4876 scope.go:117] "RemoveContainer" containerID="1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.842886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.842971 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.842989 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.843014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.843031 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.946405 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.946478 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.946496 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.946522 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:18 crc kubenswrapper[4876]: I1205 08:16:18.946539 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:18Z","lastTransitionTime":"2025-12-05T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.048788 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.049242 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.049262 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.049282 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.049294 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.100239 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/1.log" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.105425 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.106598 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.130761 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.150121 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.153257 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.153306 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.153320 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.153340 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.153354 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.168353 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.184468 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.200636 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.218171 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.241206 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.255710 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.255762 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.255776 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.255793 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.255808 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.259234 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.273556 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.287413 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.304514 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.326231 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.337319 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.346858 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.357362 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.358644 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.358705 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.358729 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.358760 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.358782 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.371680 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:19Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.462085 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.462175 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.462189 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.462210 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.462224 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.564623 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.564674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.564704 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.564725 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.564736 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.667543 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.667586 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.667597 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.667616 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.667627 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.770479 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.770528 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.770540 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.770559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.770570 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.823696 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:19 crc kubenswrapper[4876]: E1205 08:16:19.823826 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.823882 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:19 crc kubenswrapper[4876]: E1205 08:16:19.824043 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.824529 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:19 crc kubenswrapper[4876]: E1205 08:16:19.824778 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.876592 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.876645 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.876837 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.876859 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.876873 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.980750 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.980816 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.980835 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.980861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:19 crc kubenswrapper[4876]: I1205 08:16:19.980879 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:19Z","lastTransitionTime":"2025-12-05T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.083115 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.083167 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.083182 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.083202 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.083217 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.111057 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/2.log" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.111945 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/1.log" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.115617 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6" exitCode=1 Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.115669 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.115725 4876 scope.go:117] "RemoveContainer" containerID="1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.116674 4876 scope.go:117] "RemoveContainer" containerID="0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6" Dec 05 08:16:20 crc kubenswrapper[4876]: E1205 08:16:20.116960 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\"" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.143454 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.165353 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.182841 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.185854 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.185924 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.185942 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.185965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.185982 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.198615 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.215039 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.230591 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.247405 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.257573 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.268924 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.284085 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.287742 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.287773 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.287781 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.287794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.287805 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.296365 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.306837 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.317391 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.329472 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.341876 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.368471 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1026a82e89e5d082e8d2eff32698e9d0fe88c11806c2e0930845e59a73eadf50\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:06Z\\\",\\\"message\\\":\\\":06.069766 6297 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:06.069778 6297 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 08:16:06.069797 6297 factory.go:656] Stopping watch factory\\\\nI1205 08:16:06.069815 6297 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 08:16:06.069830 6297 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 08:16:06.069821 6297 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.069988 6297 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1205 08:16:06.070005 6297 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1205 08:16:06.070012 6297 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:06.069776 6297 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 08:16:06.070451 6297 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1205 08:16:06.070567 6297 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:20Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.390623 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.390688 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.390706 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.390727 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.390743 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.493339 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.493392 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.493410 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.493434 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.493455 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.595613 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.595713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.595731 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.595757 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.595773 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.698846 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.698958 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.698978 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.699004 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.699022 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.802016 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.802083 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.802099 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.802128 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.802165 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.823280 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:20 crc kubenswrapper[4876]: E1205 08:16:20.823484 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.905092 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.905171 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.905195 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.905226 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:20 crc kubenswrapper[4876]: I1205 08:16:20.905252 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:20Z","lastTransitionTime":"2025-12-05T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.008409 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.008476 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.008493 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.008517 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.008538 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.112686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.112767 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.112791 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.112822 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.112844 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.121489 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/2.log" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.126721 4876 scope.go:117] "RemoveContainer" containerID="0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6" Dec 05 08:16:21 crc kubenswrapper[4876]: E1205 08:16:21.126997 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\"" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.148013 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.168770 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.200648 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.215835 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.215965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.215995 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.216027 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.216050 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.229091 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.247442 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.265365 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.279319 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.294150 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.312768 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.319056 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.319103 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.319120 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.319142 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.319158 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.329223 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.345879 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.357698 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.371075 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.390082 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.403841 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.416065 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:21Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.421690 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.421724 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.421735 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.421754 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.421766 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.525071 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.525174 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.525197 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.525221 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.525239 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.627855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.627980 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.628001 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.628027 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.628049 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.731336 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.731367 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.731383 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.731397 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.731407 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.823452 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.823452 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.823726 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:21 crc kubenswrapper[4876]: E1205 08:16:21.823796 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:21 crc kubenswrapper[4876]: E1205 08:16:21.823671 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:21 crc kubenswrapper[4876]: E1205 08:16:21.823878 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.834117 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.834183 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.834201 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.834230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.834249 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.937014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.937077 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.937094 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.937117 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:21 crc kubenswrapper[4876]: I1205 08:16:21.937139 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:21Z","lastTransitionTime":"2025-12-05T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.041061 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.041119 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.041135 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.041160 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.041178 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.160323 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.160406 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.160420 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.160441 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.160481 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.262251 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.262296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.262308 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.262325 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.262337 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.364845 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.364923 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.364942 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.364964 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.364981 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.467834 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.468299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.468486 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.468755 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.468982 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.572079 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.573024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.573418 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.573673 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.574305 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.677089 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.677371 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.677452 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.677530 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.677603 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.780651 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.780969 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.781133 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.781275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.781419 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.823689 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:22 crc kubenswrapper[4876]: E1205 08:16:22.823887 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.884628 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.885030 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.885211 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.885390 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.885608 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.990155 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.990243 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.990287 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.990325 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:22 crc kubenswrapper[4876]: I1205 08:16:22.990351 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:22Z","lastTransitionTime":"2025-12-05T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.093132 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.093186 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.093203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.093226 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.093243 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.196171 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.196252 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.196268 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.196286 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.196298 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.298888 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.298977 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.298995 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.299019 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.299037 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.401351 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.401412 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.401431 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.401454 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.401474 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.504417 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.504495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.504518 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.504552 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.504649 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.607843 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.607882 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.607894 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.607927 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.607940 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.711531 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.711594 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.711610 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.711637 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.711655 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.814485 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.814537 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.814550 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.814570 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.814585 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.823859 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:23 crc kubenswrapper[4876]: E1205 08:16:23.823977 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.823861 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.824014 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:23 crc kubenswrapper[4876]: E1205 08:16:23.824094 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:23 crc kubenswrapper[4876]: E1205 08:16:23.824159 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.917165 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.917221 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.917242 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.917264 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:23 crc kubenswrapper[4876]: I1205 08:16:23.917280 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:23Z","lastTransitionTime":"2025-12-05T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.019597 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.019639 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.019650 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.019664 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.019676 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.124082 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.124483 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.124650 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.124776 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.124954 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.227060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.227115 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.227134 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.227158 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.227175 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.278327 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:24 crc kubenswrapper[4876]: E1205 08:16:24.278561 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:24 crc kubenswrapper[4876]: E1205 08:16:24.278642 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs podName:79225fe3-5793-4316-92da-c227edb92d53 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:40.278620641 +0000 UTC m=+64.767285303 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs") pod "network-metrics-daemon-7wg59" (UID: "79225fe3-5793-4316-92da-c227edb92d53") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.329367 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.329589 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.329674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.329760 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.329842 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.432283 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.432358 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.432376 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.432400 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.432419 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.534578 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.534784 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.534859 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.534965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.535095 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.637803 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.638349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.638430 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.638511 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.638590 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.741829 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.742016 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.742121 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.742217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.742306 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.823598 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:24 crc kubenswrapper[4876]: E1205 08:16:24.823712 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.845212 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.845273 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.845296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.845323 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.845340 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.947808 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.947934 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.947972 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.948010 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:24 crc kubenswrapper[4876]: I1205 08:16:24.948035 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:24Z","lastTransitionTime":"2025-12-05T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.050193 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.050384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.050416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.050444 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.050469 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.159189 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.159292 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.159311 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.159340 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.159363 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.262885 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.262978 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.262995 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.263017 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.263034 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.366640 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.366683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.366692 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.366708 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.366718 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.469217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.469296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.469320 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.469350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.469367 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.572337 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.572398 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.572418 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.572443 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.572460 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.622158 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.637125 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.646888 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.667022 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.675268 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.675371 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.675391 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.675447 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.675467 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.689288 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.707515 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.726718 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.743759 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.770569 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.778460 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.778561 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.778582 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.778643 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.778663 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.792498 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.809177 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.822601 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.823174 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.823174 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:25 crc kubenswrapper[4876]: E1205 08:16:25.823524 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:25 crc kubenswrapper[4876]: E1205 08:16:25.823373 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.823201 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:25 crc kubenswrapper[4876]: E1205 08:16:25.823697 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.842076 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.862607 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.883336 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.883381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.883401 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.883433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.883456 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.887949 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.905031 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.922194 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.935664 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.957878 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.978865 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.987772 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.987830 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.987851 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.987959 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.987980 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:25Z","lastTransitionTime":"2025-12-05T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:25 crc kubenswrapper[4876]: I1205 08:16:25.996530 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:25Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.018521 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.039283 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.060789 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.079778 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.091081 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.091126 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.091144 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.091170 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.091192 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.102835 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.122170 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.144452 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.165122 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.185246 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.194445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.194543 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.194568 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.194600 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.194665 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.203086 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.221570 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.243092 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.262832 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.292035 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:26Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.297537 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.297595 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.297607 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.297627 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.297639 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.400283 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.400350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.400370 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.400395 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.400416 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.503507 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.503582 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.503613 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.503642 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.503660 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.606451 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.606508 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.606525 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.606550 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.606568 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.709494 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.709555 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.709572 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.709598 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.709617 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.812776 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.812850 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.812873 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.812940 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.812969 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.823450 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:26 crc kubenswrapper[4876]: E1205 08:16:26.823599 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.916313 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.916392 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.916417 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.916447 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:26 crc kubenswrapper[4876]: I1205 08:16:26.916470 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:26Z","lastTransitionTime":"2025-12-05T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.019384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.019432 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.019450 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.019472 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.019488 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.122668 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.122736 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.122756 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.122781 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.122801 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.225971 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.226051 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.226076 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.226107 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.226129 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.329535 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.329584 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.329601 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.329623 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.329639 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.432525 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.432595 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.432612 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.432639 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.432657 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.535696 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.535756 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.535775 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.535800 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.535819 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.621861 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.622195 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:16:59.622164905 +0000 UTC m=+84.110829557 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.639185 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.639237 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.639263 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.639310 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.639329 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.742276 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.742333 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.742350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.742374 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.742394 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.774349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.774416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.774434 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.774459 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.774477 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.794976 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:27Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.801574 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.801655 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.801679 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.801715 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.801748 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.823711 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.823791 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.823659 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.823943 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.824188 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.823931 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:27Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.824495 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.824620 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.824711 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.824811 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.824994 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:59.824961999 +0000 UTC m=+84.313626661 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.824999 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.825021 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.825098 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825035 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825139 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825106 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825195 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:59.825176304 +0000 UTC m=+84.313840966 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825274 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825297 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:59.825272507 +0000 UTC m=+84.313937169 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825305 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825331 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.825399 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 08:16:59.825374899 +0000 UTC m=+84.314039601 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.830582 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.830627 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.830645 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.830668 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.830685 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.851697 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:27Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.857472 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.857570 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.857595 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.857619 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.857637 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.878659 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:27Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.883707 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.883793 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.883819 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.883852 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.884034 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.899272 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:27Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:27 crc kubenswrapper[4876]: E1205 08:16:27.899521 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.901636 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.901683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.901697 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.901717 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:27 crc kubenswrapper[4876]: I1205 08:16:27.901732 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:27Z","lastTransitionTime":"2025-12-05T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.004860 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.004960 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.004983 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.005013 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.005036 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.107769 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.107826 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.107850 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.107884 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.107922 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.211512 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.211584 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.211604 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.211629 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.211649 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.315132 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.315201 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.315220 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.315250 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.315272 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.419052 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.419115 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.419151 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.419189 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.419212 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.522975 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.523043 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.523067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.523098 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.523121 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.626474 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.626532 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.626555 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.626581 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.626602 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.729973 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.730043 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.730068 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.730100 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.730125 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.823812 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:28 crc kubenswrapper[4876]: E1205 08:16:28.824093 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.833370 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.833454 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.833468 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.833516 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.833533 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.937790 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.937872 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.937891 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.937957 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:28 crc kubenswrapper[4876]: I1205 08:16:28.937980 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:28Z","lastTransitionTime":"2025-12-05T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.041121 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.041203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.041245 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.041277 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.041301 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.144267 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.144367 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.144384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.144407 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.144426 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.248384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.248475 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.248501 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.248528 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.248544 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.351782 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.351867 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.351891 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.351951 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.351973 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.455360 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.455431 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.455450 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.455477 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.455497 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.559218 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.559258 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.559268 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.559282 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.559292 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.662230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.662271 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.662279 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.662298 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.662308 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.765470 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.765534 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.765555 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.765579 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.765596 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.823242 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.823495 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:29 crc kubenswrapper[4876]: E1205 08:16:29.823572 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.823604 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:29 crc kubenswrapper[4876]: E1205 08:16:29.823740 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:29 crc kubenswrapper[4876]: E1205 08:16:29.823794 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.869529 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.869572 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.869584 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.869633 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.869648 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.972038 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.972077 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.972086 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.972099 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:29 crc kubenswrapper[4876]: I1205 08:16:29.972111 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:29Z","lastTransitionTime":"2025-12-05T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.075585 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.075641 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.075657 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.075681 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.075696 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.178964 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.179045 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.179072 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.179102 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.179126 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.282707 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.282782 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.282806 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.282839 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.282863 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.385344 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.385413 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.385433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.385458 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.385476 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.488605 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.488686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.488714 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.488746 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.488768 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.591054 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.591098 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.591110 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.591125 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.591135 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.694321 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.694396 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.694442 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.694473 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.694633 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.797736 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.797779 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.797792 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.797810 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.797824 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.823342 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:30 crc kubenswrapper[4876]: E1205 08:16:30.823497 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.900449 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.900521 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.900533 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.900554 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:30 crc kubenswrapper[4876]: I1205 08:16:30.900566 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:30Z","lastTransitionTime":"2025-12-05T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.003695 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.003738 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.003751 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.003768 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.003780 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.106081 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.106156 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.106174 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.106201 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.106221 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.209828 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.209893 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.209949 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.209973 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.209990 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.313209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.313275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.313293 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.313317 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.313336 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.416555 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.416625 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.416649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.416677 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.416699 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.520192 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.520258 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.520275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.520299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.520317 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.622348 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.622444 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.622476 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.622505 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.622525 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.725282 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.725347 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.725365 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.725388 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.725406 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.823699 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.823788 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.823715 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:31 crc kubenswrapper[4876]: E1205 08:16:31.823970 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:31 crc kubenswrapper[4876]: E1205 08:16:31.824137 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:31 crc kubenswrapper[4876]: E1205 08:16:31.824211 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.828271 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.828317 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.828334 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.828358 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.828375 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.931993 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.932034 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.932046 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.932060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:31 crc kubenswrapper[4876]: I1205 08:16:31.932072 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:31Z","lastTransitionTime":"2025-12-05T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.035175 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.035212 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.035221 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.035235 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.035247 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.138022 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.138064 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.138075 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.138089 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.138099 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.241708 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.241769 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.241786 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.241810 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.241829 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.344458 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.344530 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.344548 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.344572 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.344591 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.447734 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.447784 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.447800 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.447825 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.447842 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.550912 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.550951 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.550960 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.550973 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.550984 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.655227 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.655298 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.655318 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.655386 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.655413 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.757663 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.757714 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.757725 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.757743 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.757753 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.823790 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:32 crc kubenswrapper[4876]: E1205 08:16:32.824029 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.860019 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.860093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.860110 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.860138 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.860157 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.964979 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.965053 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.965084 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.965131 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:32 crc kubenswrapper[4876]: I1205 08:16:32.965159 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:32Z","lastTransitionTime":"2025-12-05T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.068186 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.068242 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.068254 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.068272 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.068287 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.170706 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.170772 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.170791 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.170815 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.170834 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.274184 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.274256 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.274273 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.274299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.274316 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.377207 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.377256 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.377273 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.377298 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.377315 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.480636 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.481119 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.481317 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.481495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.481673 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.585638 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.585710 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.585728 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.585753 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.585773 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.689266 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.689336 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.689356 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.689567 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.689586 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.792349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.792397 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.792414 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.792435 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.792451 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.824299 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:33 crc kubenswrapper[4876]: E1205 08:16:33.824517 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.824662 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:33 crc kubenswrapper[4876]: E1205 08:16:33.824823 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.825070 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:33 crc kubenswrapper[4876]: E1205 08:16:33.825385 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.894495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.894541 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.894552 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.894570 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.894583 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.997381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.997452 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.997480 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.997511 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:33 crc kubenswrapper[4876]: I1205 08:16:33.997534 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:33Z","lastTransitionTime":"2025-12-05T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.100878 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.100987 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.101012 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.101047 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.101071 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.203554 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.203683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.203702 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.203732 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.203757 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.307357 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.307422 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.307444 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.307471 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.307491 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.409967 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.410024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.410043 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.410080 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.410096 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.512492 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.512562 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.512580 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.512606 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.512626 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.615816 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.615960 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.615989 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.616024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.616048 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.719732 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.719825 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.719842 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.719865 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.719883 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.822736 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.822853 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.822878 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.822941 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.823047 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: E1205 08:16:34.823103 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.823123 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.925649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.925719 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.925744 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.925773 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:34 crc kubenswrapper[4876]: I1205 08:16:34.925793 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:34Z","lastTransitionTime":"2025-12-05T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.028976 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.029045 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.029067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.029093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.029117 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.132588 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.132735 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.132760 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.132831 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.132859 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.235306 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.235365 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.235383 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.235407 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.235427 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.337445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.337505 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.337524 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.337550 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.337572 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.440316 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.440373 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.440389 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.440411 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.440427 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.543252 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.543322 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.543344 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.543370 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.543391 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.650136 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.650184 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.650192 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.650226 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.650253 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.753353 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.753757 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.753774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.753799 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.753819 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.822894 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.823025 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.823773 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:35 crc kubenswrapper[4876]: E1205 08:16:35.824136 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:35 crc kubenswrapper[4876]: E1205 08:16:35.823976 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:35 crc kubenswrapper[4876]: E1205 08:16:35.824432 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.824736 4876 scope.go:117] "RemoveContainer" containerID="0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6" Dec 05 08:16:35 crc kubenswrapper[4876]: E1205 08:16:35.825216 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\"" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.837824 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:35Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.856446 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.856687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.856707 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.856731 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.856749 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.857867 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:35Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.892944 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:35Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.920829 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:35Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.941155 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:35Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.960664 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:35Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.960767 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.960805 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.960821 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.960843 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.960860 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:35Z","lastTransitionTime":"2025-12-05T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.977613 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:35Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:35 crc kubenswrapper[4876]: I1205 08:16:35.991934 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:35Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.004824 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.020121 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.039170 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.055423 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.063849 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.063941 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.063966 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.063994 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.064014 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.072200 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.087069 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.107768 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.125230 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.144889 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:36Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.165796 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.165827 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.165836 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.165851 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.165859 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.267817 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.267865 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.267878 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.267894 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.267930 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.371068 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.371129 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.371149 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.371173 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.371191 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.474158 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.474222 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.474243 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.474275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.474296 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.577058 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.577110 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.577130 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.577154 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.577170 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.680499 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.680547 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.680564 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.680587 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.680604 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.783354 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.783436 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.783458 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.783482 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.783499 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.823673 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:36 crc kubenswrapper[4876]: E1205 08:16:36.823866 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.887562 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.887619 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.887640 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.887664 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.887681 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.990463 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.990494 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.990507 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.990519 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:36 crc kubenswrapper[4876]: I1205 08:16:36.990527 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:36Z","lastTransitionTime":"2025-12-05T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.093677 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.093763 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.093786 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.093817 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.093840 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.196066 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.196128 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.196144 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.196166 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.196184 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.299338 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.299391 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.299407 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.299431 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.299449 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.402968 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.403039 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.403060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.403089 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.403111 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.506669 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.506940 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.506981 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.507013 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.507036 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.609080 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.609137 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.609150 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.609167 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.609178 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.713874 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.714586 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.714608 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.714629 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.714645 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.817318 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.817375 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.817392 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.817416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.817437 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.823719 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.823827 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.823963 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:37 crc kubenswrapper[4876]: E1205 08:16:37.823988 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:37 crc kubenswrapper[4876]: E1205 08:16:37.824120 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:37 crc kubenswrapper[4876]: E1205 08:16:37.824202 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.921154 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.921207 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.921219 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.921236 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:37 crc kubenswrapper[4876]: I1205 08:16:37.921250 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:37Z","lastTransitionTime":"2025-12-05T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.023886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.023952 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.023964 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.023984 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.023995 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.126976 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.127021 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.127033 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.127050 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.127064 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.211678 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.211746 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.211763 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.211802 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.211819 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: E1205 08:16:38.232993 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:38Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.237960 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.238022 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.238037 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.238057 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.238071 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: E1205 08:16:38.255718 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:38Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.259662 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.259698 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.259716 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.259733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.259747 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: E1205 08:16:38.277536 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:38Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.281745 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.281791 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.281809 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.281830 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.281845 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: E1205 08:16:38.300848 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:38Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.305204 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.305238 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.305248 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.305266 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.305285 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: E1205 08:16:38.322012 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:38Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:38 crc kubenswrapper[4876]: E1205 08:16:38.322157 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.323883 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.323979 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.323997 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.324022 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.324039 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.426988 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.427050 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.427073 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.427100 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.427121 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.529674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.529738 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.529759 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.529787 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.529811 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.632371 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.632433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.632451 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.632482 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.632503 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.735838 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.735881 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.735893 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.735924 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.735939 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.823384 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:38 crc kubenswrapper[4876]: E1205 08:16:38.823609 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.838662 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.838743 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.838762 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.838790 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.838812 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.942708 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.942774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.942794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.942817 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:38 crc kubenswrapper[4876]: I1205 08:16:38.942835 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:38Z","lastTransitionTime":"2025-12-05T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.045950 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.046006 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.046024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.046050 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.046070 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.148440 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.148480 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.148508 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.148523 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.148536 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.251209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.251260 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.251272 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.251291 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.251305 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.353950 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.353982 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.353991 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.354006 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.354017 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.456668 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.456730 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.456752 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.456783 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.456805 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.559038 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.559075 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.559084 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.559098 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.559106 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.661700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.661735 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.661744 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.661757 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.661766 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.764207 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.764246 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.764259 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.764275 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.764290 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.824159 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.824205 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:39 crc kubenswrapper[4876]: E1205 08:16:39.824257 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.824159 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:39 crc kubenswrapper[4876]: E1205 08:16:39.824429 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:39 crc kubenswrapper[4876]: E1205 08:16:39.824528 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.866656 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.866700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.866711 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.866727 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.866739 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.969209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.969239 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.969247 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.969261 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:39 crc kubenswrapper[4876]: I1205 08:16:39.969270 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:39Z","lastTransitionTime":"2025-12-05T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.071649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.071673 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.071681 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.071692 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.071701 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.174190 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.174221 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.174228 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.174242 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.174251 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.277044 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.277074 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.277082 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.277095 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.277106 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.295383 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:40 crc kubenswrapper[4876]: E1205 08:16:40.295498 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:40 crc kubenswrapper[4876]: E1205 08:16:40.295543 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs podName:79225fe3-5793-4316-92da-c227edb92d53 nodeName:}" failed. No retries permitted until 2025-12-05 08:17:12.295530909 +0000 UTC m=+96.784195531 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs") pod "network-metrics-daemon-7wg59" (UID: "79225fe3-5793-4316-92da-c227edb92d53") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.380189 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.380217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.380226 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.380239 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.380248 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.482541 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.482592 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.482601 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.482616 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.482627 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.585063 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.585123 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.585143 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.585170 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.585188 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.687765 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.687830 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.687847 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.687870 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.687886 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.790790 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.790828 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.790838 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.790854 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.790863 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.823408 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:40 crc kubenswrapper[4876]: E1205 08:16:40.823538 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.893702 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.893740 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.893749 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.893764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.893773 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.997802 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.997853 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.997871 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.997894 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:40 crc kubenswrapper[4876]: I1205 08:16:40.997955 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:40Z","lastTransitionTime":"2025-12-05T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.100753 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.100784 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.100794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.100810 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.100820 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.203439 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.203475 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.203487 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.203504 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.203514 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.307175 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.307217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.307228 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.307244 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.307255 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.409557 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.409590 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.409597 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.409611 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.409620 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.512007 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.512045 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.512054 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.512067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.512075 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.614603 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.614647 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.614660 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.614702 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.614714 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.717669 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.717706 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.717716 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.717729 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.717738 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.819803 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.819829 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.819842 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.819856 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.819867 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.823279 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.823305 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.823305 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:41 crc kubenswrapper[4876]: E1205 08:16:41.823456 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:41 crc kubenswrapper[4876]: E1205 08:16:41.823526 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:41 crc kubenswrapper[4876]: E1205 08:16:41.823624 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.922565 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.922605 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.922613 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.922626 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:41 crc kubenswrapper[4876]: I1205 08:16:41.922634 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:41Z","lastTransitionTime":"2025-12-05T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.026304 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.026370 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.026720 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.027035 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.027089 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.129719 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.129751 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.129759 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.129774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.129784 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.231673 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.231726 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.231745 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.231770 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.231789 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.334553 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.334595 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.334608 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.334627 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.334639 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.438014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.438091 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.438119 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.438152 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.438175 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.540678 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.540794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.540814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.540841 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.540864 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.644331 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.644392 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.644408 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.644427 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.644440 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.746741 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.746796 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.746814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.746836 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.746852 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.823621 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:42 crc kubenswrapper[4876]: E1205 08:16:42.823759 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.849430 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.849492 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.849512 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.849537 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.849555 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.951857 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.951963 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.951986 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.952015 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:42 crc kubenswrapper[4876]: I1205 08:16:42.952038 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:42Z","lastTransitionTime":"2025-12-05T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.055153 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.055205 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.055217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.055231 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.055241 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.157124 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.157191 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.157207 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.157233 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.157252 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.204589 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/0.log" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.204638 4876 generic.go:334] "Generic (PLEG): container finished" podID="e7274e0d-7f30-495c-800a-5dcbf1656b0d" containerID="3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc" exitCode=1 Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.204664 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jwrfx" event={"ID":"e7274e0d-7f30-495c-800a-5dcbf1656b0d","Type":"ContainerDied","Data":"3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.205031 4876 scope.go:117] "RemoveContainer" containerID="3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.220459 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.235260 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:42Z\\\",\\\"message\\\":\\\"2025-12-05T08:15:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b\\\\n2025-12-05T08:15:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b to /host/opt/cni/bin/\\\\n2025-12-05T08:15:57Z [verbose] multus-daemon started\\\\n2025-12-05T08:15:57Z [verbose] Readiness Indicator file check\\\\n2025-12-05T08:16:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.249025 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.259035 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.259106 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.259121 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.259138 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.259151 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.266682 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.280413 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.293866 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.306489 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.316635 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.326214 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.338814 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.351814 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.361768 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.361823 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.361839 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.361861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.361876 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.366136 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.379204 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.391689 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.406823 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.420112 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.437654 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:43Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.464251 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.464294 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.464306 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.464323 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.464335 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.566469 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.566509 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.566519 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.566534 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.566544 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.669080 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.669134 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.669151 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.669170 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.669184 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.772482 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.772554 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.772576 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.772607 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.772628 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.823850 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.823863 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:43 crc kubenswrapper[4876]: E1205 08:16:43.824150 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.823884 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:43 crc kubenswrapper[4876]: E1205 08:16:43.824372 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:43 crc kubenswrapper[4876]: E1205 08:16:43.824440 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.875328 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.875385 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.875403 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.875425 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.875439 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.979871 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.979920 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.979932 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.979949 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:43 crc kubenswrapper[4876]: I1205 08:16:43.979960 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:43Z","lastTransitionTime":"2025-12-05T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.081984 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.082024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.082034 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.082048 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.082059 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.185004 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.185047 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.185056 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.185068 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.185077 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.209653 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/0.log" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.209710 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jwrfx" event={"ID":"e7274e0d-7f30-495c-800a-5dcbf1656b0d","Type":"ContainerStarted","Data":"bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.229030 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.242139 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.258739 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.268408 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.278190 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.286816 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.286865 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.286875 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.286889 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.286933 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.288377 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.304065 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.323160 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.335341 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.345504 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.357440 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.377135 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.387354 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.388645 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.388672 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.388683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.388699 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.388710 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.397888 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.407360 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.421075 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.432641 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:42Z\\\",\\\"message\\\":\\\"2025-12-05T08:15:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b\\\\n2025-12-05T08:15:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b to /host/opt/cni/bin/\\\\n2025-12-05T08:15:57Z [verbose] multus-daemon started\\\\n2025-12-05T08:15:57Z [verbose] Readiness Indicator file check\\\\n2025-12-05T08:16:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:44Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.490624 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.490669 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.490682 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.490700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.490711 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.593554 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.593623 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.593635 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.593654 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.593662 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.695759 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.695794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.695802 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.695821 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.695832 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.797772 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.797811 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.797822 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.797837 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.797849 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.823519 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:44 crc kubenswrapper[4876]: E1205 08:16:44.823734 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.900863 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.900938 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.900946 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.900961 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:44 crc kubenswrapper[4876]: I1205 08:16:44.900970 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:44Z","lastTransitionTime":"2025-12-05T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.003725 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.003767 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.003779 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.003794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.003806 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.106064 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.106140 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.106166 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.106192 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.106210 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.208894 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.208974 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.208990 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.209016 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.209039 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.312194 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.312273 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.312296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.312321 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.312341 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.414475 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.414516 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.414527 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.414543 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.414554 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.516837 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.516878 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.516887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.516917 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.516931 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.618830 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.618868 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.618879 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.618909 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.618925 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.720955 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.720992 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.721005 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.721024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.721036 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.822836 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.823019 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.823029 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:45 crc kubenswrapper[4876]: E1205 08:16:45.823182 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.823343 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.823408 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.823425 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.823489 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.823507 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: E1205 08:16:45.823993 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:45 crc kubenswrapper[4876]: E1205 08:16:45.824088 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.832299 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.843939 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.857284 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:42Z\\\",\\\"message\\\":\\\"2025-12-05T08:15:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b\\\\n2025-12-05T08:15:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b to /host/opt/cni/bin/\\\\n2025-12-05T08:15:57Z [verbose] multus-daemon started\\\\n2025-12-05T08:15:57Z [verbose] Readiness Indicator file check\\\\n2025-12-05T08:16:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.871731 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.887113 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.905950 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.919095 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.925531 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.925564 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.925575 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.925589 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.925598 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:45Z","lastTransitionTime":"2025-12-05T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.930661 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.945634 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.958365 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.975003 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:45 crc kubenswrapper[4876]: I1205 08:16:45.990089 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:45Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.006183 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:46Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.020168 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:46Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.028092 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.028180 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.028203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.028228 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.028245 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.034697 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:46Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.049060 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:46Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.061809 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:46Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.080985 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:46Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.132125 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.132172 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.132182 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.132199 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.132211 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.234219 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.234255 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.234264 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.234280 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.234292 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.337019 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.337079 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.337098 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.337124 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.337142 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.439173 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.439244 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.439267 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.439295 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.439316 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.541538 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.541589 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.541601 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.541619 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.541632 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.644641 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.644696 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.644714 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.644738 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.644755 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.747378 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.747418 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.747435 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.747455 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.747471 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.823455 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:46 crc kubenswrapper[4876]: E1205 08:16:46.823631 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.849958 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.850014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.850036 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.850063 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.850086 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.951866 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.951913 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.951926 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.951937 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:46 crc kubenswrapper[4876]: I1205 08:16:46.951946 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:46Z","lastTransitionTime":"2025-12-05T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.055310 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.055363 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.055381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.055403 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.055420 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.158520 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.158561 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.158569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.158584 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.158592 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.261082 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.261143 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.261160 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.261221 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.261240 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.363613 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.363661 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.363679 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.363702 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.363719 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.466861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.466918 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.466932 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.466945 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.466954 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.569209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.569270 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.569286 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.569324 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.569341 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.672162 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.672203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.672215 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.672229 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.672239 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.774675 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.774701 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.774712 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.774724 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.774734 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.823192 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:47 crc kubenswrapper[4876]: E1205 08:16:47.823279 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.823343 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.823365 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:47 crc kubenswrapper[4876]: E1205 08:16:47.823544 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:47 crc kubenswrapper[4876]: E1205 08:16:47.823635 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.876890 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.876948 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.876959 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.876975 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.876986 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.979931 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.979992 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.980010 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.980036 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:47 crc kubenswrapper[4876]: I1205 08:16:47.980053 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:47Z","lastTransitionTime":"2025-12-05T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.083674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.083728 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.083743 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.083765 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.083779 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.186688 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.186764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.186782 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.186808 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.186825 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.290283 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.290350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.290375 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.290404 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.290421 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.375754 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.375811 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.375827 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.375852 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.375873 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: E1205 08:16:48.394127 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:48Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.400636 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.400733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.400747 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.400764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.400778 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: E1205 08:16:48.417747 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:48Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.421532 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.421558 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.421569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.421586 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.421597 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: E1205 08:16:48.437315 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:48Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.441707 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.441739 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.441749 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.441764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.441776 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: E1205 08:16:48.456511 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:48Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.461654 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.461682 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.461694 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.461711 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.461723 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: E1205 08:16:48.479701 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:48Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:48 crc kubenswrapper[4876]: E1205 08:16:48.479844 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.481381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.481442 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.481460 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.481487 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.481504 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.584651 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.584709 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.584728 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.584750 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.584768 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.687082 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.687116 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.687127 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.687145 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.687156 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.790155 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.790228 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.790246 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.790271 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.790289 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.823738 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:48 crc kubenswrapper[4876]: E1205 08:16:48.823997 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.825145 4876 scope.go:117] "RemoveContainer" containerID="0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.892353 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.892409 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.892428 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.892463 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.892481 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.995349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.995417 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.995440 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.995470 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:48 crc kubenswrapper[4876]: I1205 08:16:48.995495 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:48Z","lastTransitionTime":"2025-12-05T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.098511 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.098576 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.098596 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.098619 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.098636 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.200972 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.201005 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.201017 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.201034 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.201045 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.228639 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/2.log" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.233087 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.233806 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.256083 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.275606 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.296406 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.307601 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.307679 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.307703 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.307733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.307755 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.314990 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.328744 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.341976 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c5fa857-dcef-43e2-8f42-0b94fcd00541\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a056e141e0c391539ffd94a814372c68901b5e7ea03c7c696c333c5fa320142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.355991 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.376352 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.401968 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.410847 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.410922 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.410940 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.410966 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.411011 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.420592 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.442446 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.465587 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.489994 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.517813 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.517865 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.517887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.517932 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.517949 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.519287 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.539390 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.558169 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.582534 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.603998 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:42Z\\\",\\\"message\\\":\\\"2025-12-05T08:15:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b\\\\n2025-12-05T08:15:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b to /host/opt/cni/bin/\\\\n2025-12-05T08:15:57Z [verbose] multus-daemon started\\\\n2025-12-05T08:15:57Z [verbose] Readiness Indicator file check\\\\n2025-12-05T08:16:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:49Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.620849 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.620941 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.620962 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.620987 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.621005 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.724267 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.724318 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.724334 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.724358 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.724375 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.823127 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.823211 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:49 crc kubenswrapper[4876]: E1205 08:16:49.823328 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.823220 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:49 crc kubenswrapper[4876]: E1205 08:16:49.823438 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:49 crc kubenswrapper[4876]: E1205 08:16:49.823508 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.827589 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.827645 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.827663 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.827684 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.827701 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.931290 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.931416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.931435 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.932055 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:49 crc kubenswrapper[4876]: I1205 08:16:49.932142 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:49Z","lastTransitionTime":"2025-12-05T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.035935 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.035997 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.036014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.036036 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.036053 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.139369 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.139473 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.139493 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.139517 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.139536 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.242469 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.242518 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.242536 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.242557 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.242576 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.345810 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.345871 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.345890 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.345942 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.345964 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.449389 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.449447 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.449464 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.449491 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.449509 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.552686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.552736 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.552753 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.552778 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.552795 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.656775 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.656825 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.656843 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.656866 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.656885 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.759786 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.759860 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.760072 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.760100 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.760119 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.823997 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:50 crc kubenswrapper[4876]: E1205 08:16:50.824203 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.863358 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.863407 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.863418 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.863434 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.863446 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.967569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.967622 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.967642 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.967665 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:50 crc kubenswrapper[4876]: I1205 08:16:50.967682 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:50Z","lastTransitionTime":"2025-12-05T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.070724 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.070783 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.070803 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.070827 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.070845 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.174366 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.174484 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.174510 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.174538 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.174558 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.247311 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/3.log" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.248293 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/2.log" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.252116 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1" exitCode=1 Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.252172 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.252253 4876 scope.go:117] "RemoveContainer" containerID="0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.253413 4876 scope.go:117] "RemoveContainer" containerID="f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1" Dec 05 08:16:51 crc kubenswrapper[4876]: E1205 08:16:51.253697 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\"" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.277404 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.277451 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.277463 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.277482 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.277494 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.278001 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.297882 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:42Z\\\",\\\"message\\\":\\\"2025-12-05T08:15:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b\\\\n2025-12-05T08:15:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b to /host/opt/cni/bin/\\\\n2025-12-05T08:15:57Z [verbose] multus-daemon started\\\\n2025-12-05T08:15:57Z [verbose] Readiness Indicator file check\\\\n2025-12-05T08:16:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.314419 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.334024 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.348634 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.363531 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.380567 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.381442 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.381531 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.381554 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.381589 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.381613 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.399477 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c5fa857-dcef-43e2-8f42-0b94fcd00541\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a056e141e0c391539ffd94a814372c68901b5e7ea03c7c696c333c5fa320142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.423265 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.445345 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.463302 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.481378 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.485400 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.485443 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.485456 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.485478 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.485494 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.499448 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.519699 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.536840 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.553163 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.569759 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.587776 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.587817 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.587830 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.587845 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.587857 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.592788 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:50Z\\\",\\\"message\\\":\\\"rvices.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/community-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"d389393c-7ba9-422c-b3f5-06e391d537d2\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/community-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.189\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1205 08:16:50.071429 6853 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:51Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.690231 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.690278 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.690294 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.690313 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.690329 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.792777 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.792825 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.792837 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.792855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.792867 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.823682 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.823729 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.823768 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:51 crc kubenswrapper[4876]: E1205 08:16:51.823812 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:51 crc kubenswrapper[4876]: E1205 08:16:51.823913 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:51 crc kubenswrapper[4876]: E1205 08:16:51.824106 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.895535 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.895575 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.895588 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.895602 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.895614 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.998297 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.998378 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.998406 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.998438 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:51 crc kubenswrapper[4876]: I1205 08:16:51.998462 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:51Z","lastTransitionTime":"2025-12-05T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.101388 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.101434 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.101445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.101462 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.101474 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.204868 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.205315 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.205332 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.205356 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.205373 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.257792 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/3.log" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.307511 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.307548 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.307564 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.307585 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.307596 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.410713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.410763 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.410774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.410794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.410806 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.514149 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.514187 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.514199 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.514216 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.514228 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.617240 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.617293 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.617310 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.617333 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.617351 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.720416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.720540 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.720559 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.720582 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.720601 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.822812 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:52 crc kubenswrapper[4876]: E1205 08:16:52.823070 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.823703 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.823745 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.823757 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.823774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.823787 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.927481 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.927517 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.927528 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.927543 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:52 crc kubenswrapper[4876]: I1205 08:16:52.927554 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:52Z","lastTransitionTime":"2025-12-05T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.030330 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.030452 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.030482 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.030505 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.030522 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.133922 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.133972 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.133983 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.134000 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.134011 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.237298 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.237367 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.237458 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.237492 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.237555 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.344500 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.344605 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.344628 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.344662 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.344697 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.449067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.449137 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.449156 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.449180 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.449198 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.552851 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.552949 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.552974 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.553002 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.553072 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.656079 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.656144 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.656164 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.656189 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.656207 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.759348 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.759407 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.759429 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.759458 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.759481 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.823036 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:53 crc kubenswrapper[4876]: E1205 08:16:53.823211 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.823470 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:53 crc kubenswrapper[4876]: E1205 08:16:53.823568 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.823771 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:53 crc kubenswrapper[4876]: E1205 08:16:53.823876 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.862939 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.863390 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.863433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.863463 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.863483 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.966155 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.966198 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.966210 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.966228 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:53 crc kubenswrapper[4876]: I1205 08:16:53.966239 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:53Z","lastTransitionTime":"2025-12-05T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.069329 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.069374 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.069391 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.069413 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.069430 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.172948 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.173032 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.173057 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.173086 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.173110 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.276101 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.276179 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.276198 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.276224 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.276243 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.379830 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.379949 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.379975 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.380005 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.380032 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.483314 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.483384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.483401 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.483428 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.483448 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.586729 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.586856 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.586881 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.586945 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.586971 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.689525 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.689563 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.689571 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.689590 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.689600 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.792492 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.792546 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.792565 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.792591 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.792608 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.825358 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:54 crc kubenswrapper[4876]: E1205 08:16:54.825546 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.895434 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.895500 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.895520 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.895545 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.895562 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.998355 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.998477 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.998491 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.998508 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:54 crc kubenswrapper[4876]: I1205 08:16:54.998520 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:54Z","lastTransitionTime":"2025-12-05T08:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.100707 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.100749 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.100756 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.100772 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.100781 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.203422 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.203530 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.203590 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.203618 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.203640 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.306533 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.306604 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.306622 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.306647 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.306665 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.409112 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.409181 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.409198 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.409222 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.409240 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.512649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.512697 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.512708 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.512724 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.512737 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.615807 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.615847 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.615855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.615869 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.615877 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.718317 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.718361 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.718373 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.718389 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.718401 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.822484 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.822541 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.822553 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.822571 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.822583 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.822797 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.822826 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:55 crc kubenswrapper[4876]: E1205 08:16:55.822889 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.822921 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:55 crc kubenswrapper[4876]: E1205 08:16:55.823007 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:55 crc kubenswrapper[4876]: E1205 08:16:55.823127 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.841651 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.858960 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.877329 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e039fe6151d8f6ccaaeaee614d300cad6d643dbfed938b0965a630f90c791c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:20Z\\\",\\\"message\\\":\\\"[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.1:443:192.168.126.11:6443]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {b21188fe-5483-4717-afe6-20a41a40b91a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 08:16:19.896430 6502 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1205 08:16:19.896438 6502 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1205 08:16:19.896458 6502 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1205 08:16:19.896502 6502 factory.go:656] Stopping watch factory\\\\nI1205 08:16:19.896522 6502 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1205 08:16:19.896535 6502 handler.go:208] Removed *v1.Node event handler 2\\\\nI1205 08:16:19.896544 6502 handler.go:208] Removed *v1.Node event handler 7\\\\nI1205 08:16:19.896730 6502 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1205 08:16:19.896795 6502 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1205 08:16:19.896826 6502 ovnkube.go:599] Stopped ovnkube\\\\nI1205 08:16:19.896858 6502 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1205 08:16:19.896930 6502 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:50Z\\\",\\\"message\\\":\\\"rvices.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/community-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"d389393c-7ba9-422c-b3f5-06e391d537d2\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/community-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.189\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1205 08:16:50.071429 6853 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.893166 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.912939 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:42Z\\\",\\\"message\\\":\\\"2025-12-05T08:15:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b\\\\n2025-12-05T08:15:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b to /host/opt/cni/bin/\\\\n2025-12-05T08:15:57Z [verbose] multus-daemon started\\\\n2025-12-05T08:15:57Z [verbose] Readiness Indicator file check\\\\n2025-12-05T08:16:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.924920 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.924956 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.924970 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.924987 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.924999 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:55Z","lastTransitionTime":"2025-12-05T08:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.930306 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.945278 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c5fa857-dcef-43e2-8f42-0b94fcd00541\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a056e141e0c391539ffd94a814372c68901b5e7ea03c7c696c333c5fa320142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.963835 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.978300 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:55 crc kubenswrapper[4876]: I1205 08:16:55.994341 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:55Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.007542 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.019500 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.027245 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.027283 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.027328 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.027345 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.027357 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.035298 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.052703 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.064017 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.075592 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.086156 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.099857 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:56Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.129643 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.129688 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.129701 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.129719 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.129731 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.232654 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.232733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.232752 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.232781 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.232833 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.335875 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.335978 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.336004 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.336030 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.336048 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.438434 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.438479 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.438490 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.438506 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.438519 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.542457 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.542543 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.542569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.542600 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.542624 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.645372 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.645425 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.645443 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.645468 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.645485 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.750023 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.750091 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.750108 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.750133 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.750150 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.823874 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:56 crc kubenswrapper[4876]: E1205 08:16:56.824270 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.854749 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.854983 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.855039 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.855070 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.855088 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.958791 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.959222 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.959367 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.959505 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:56 crc kubenswrapper[4876]: I1205 08:16:56.959662 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:56Z","lastTransitionTime":"2025-12-05T08:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.062534 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.062962 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.063140 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.063333 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.063505 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.166235 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.166302 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.166320 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.166346 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.166364 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.268329 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.268371 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.268385 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.268405 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.268419 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.375759 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.376085 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.376195 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.376288 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.376369 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.479014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.479075 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.479096 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.479124 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.479146 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.581994 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.582039 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.582052 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.582068 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.582080 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.684258 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.684304 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.684315 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.684330 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.684342 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.786864 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.787174 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.787340 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.787496 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.787584 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.823655 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.823655 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.823748 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:57 crc kubenswrapper[4876]: E1205 08:16:57.824185 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:57 crc kubenswrapper[4876]: E1205 08:16:57.824261 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:57 crc kubenswrapper[4876]: E1205 08:16:57.824329 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.890256 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.890296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.890307 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.890322 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.890333 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.993162 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.993206 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.993219 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.993235 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:57 crc kubenswrapper[4876]: I1205 08:16:57.993247 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:57Z","lastTransitionTime":"2025-12-05T08:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.095680 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.095728 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.095743 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.095763 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.095777 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.199414 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.199478 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.199496 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.199531 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.199549 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.302852 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.303218 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.303239 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.303266 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.303285 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.407316 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.407382 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.407402 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.407430 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.407455 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.509474 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.509514 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.509526 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.509542 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.509554 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.609325 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.609358 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.609367 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.609380 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.609387 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: E1205 08:16:58.623475 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.627714 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.627763 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.627776 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.627794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.627806 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: E1205 08:16:58.648039 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.652146 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.652174 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.652186 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.652203 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.652215 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: E1205 08:16:58.669415 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.673060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.673102 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.673111 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.673124 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.673134 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: E1205 08:16:58.684930 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.687895 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.687948 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.687961 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.687979 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.687991 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: E1205 08:16:58.699211 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:16:58Z is after 2025-08-24T17:21:41Z" Dec 05 08:16:58 crc kubenswrapper[4876]: E1205 08:16:58.699353 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.700805 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.700833 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.700841 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.700852 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.700859 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.803553 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.803588 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.803600 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.803617 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.803629 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.823142 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:58 crc kubenswrapper[4876]: E1205 08:16:58.823316 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.906830 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.906937 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.906957 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.906979 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:58 crc kubenswrapper[4876]: I1205 08:16:58.906996 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:58Z","lastTransitionTime":"2025-12-05T08:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.009841 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.009887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.009920 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.009936 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.009947 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.112888 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.113001 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.113023 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.113048 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.113066 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.216350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.216443 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.216499 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.216536 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.216564 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.319454 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.319536 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.319554 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.319577 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.319596 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.423060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.423150 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.423178 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.423207 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.423229 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.527391 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.527452 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.527521 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.527547 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.527561 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.629815 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.629854 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.629862 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.629876 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.629886 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.696729 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.697021 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:18:03.696986113 +0000 UTC m=+148.185650795 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.732826 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.732892 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.732950 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.732976 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.732993 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.823609 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.823657 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.823671 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.823765 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.824012 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.824122 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.870639 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.870679 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.870694 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.870709 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.870721 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.899212 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.899268 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.899316 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.899340 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.899451 4876 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.899505 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:18:03.899490013 +0000 UTC m=+148.388154655 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.899697 4876 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.899732 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 08:18:03.899722685 +0000 UTC m=+148.388387327 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.899886 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.899942 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.899955 4876 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.899988 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 08:18:03.899977428 +0000 UTC m=+148.388642070 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.900145 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.900172 4876 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.900184 4876 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:59 crc kubenswrapper[4876]: E1205 08:16:59.900241 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 08:18:03.900221111 +0000 UTC m=+148.388885833 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.973097 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.973150 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.973162 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.973179 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:16:59 crc kubenswrapper[4876]: I1205 08:16:59.973191 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:16:59Z","lastTransitionTime":"2025-12-05T08:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.075329 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.075405 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.075433 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.075465 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.075486 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.178557 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.178639 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.178658 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.178681 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.178698 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.281517 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.281555 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.281564 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.281577 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.281586 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.384337 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.384381 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.384393 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.384408 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.384418 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.487426 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.487469 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.487479 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.487495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.487506 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.590004 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.590067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.590103 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.590134 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.590155 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.693234 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.693279 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.693290 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.693304 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.693314 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.795538 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.795575 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.795586 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.795601 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.795610 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.822854 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:00 crc kubenswrapper[4876]: E1205 08:17:00.822997 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.898879 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.898988 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.899012 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.899035 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:00 crc kubenswrapper[4876]: I1205 08:17:00.899051 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:00Z","lastTransitionTime":"2025-12-05T08:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.001023 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.001063 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.001072 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.001087 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.001097 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.103743 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.103812 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.103833 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.103861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.103885 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.207805 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.207875 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.207892 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.207948 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.207962 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.310129 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.310178 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.310193 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.310218 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.310234 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.413562 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.413596 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.413604 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.413617 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.413626 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.516347 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.516409 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.516426 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.516451 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.516469 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.619515 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.619580 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.619601 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.619649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.619672 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.721764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.721806 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.721814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.721829 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.721839 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.823434 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.823444 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.823484 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:01 crc kubenswrapper[4876]: E1205 08:17:01.823781 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:01 crc kubenswrapper[4876]: E1205 08:17:01.823993 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:01 crc kubenswrapper[4876]: E1205 08:17:01.823870 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.825807 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.825837 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.825846 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.825859 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.825869 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.928257 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.928289 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.928299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.928314 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:01 crc kubenswrapper[4876]: I1205 08:17:01.928323 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:01Z","lastTransitionTime":"2025-12-05T08:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.031034 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.031095 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.031112 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.031135 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.031154 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.133579 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.133660 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.133683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.133716 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.133740 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.236814 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.236888 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.236958 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.236987 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.237011 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.339325 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.339406 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.339432 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.339463 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.339485 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.442579 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.442707 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.442738 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.442766 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.442786 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.546701 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.546782 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.546808 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.546839 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.546860 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.650148 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.650218 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.650241 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.650269 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.650292 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.752988 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.753060 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.753082 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.753111 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.753132 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.823209 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:02 crc kubenswrapper[4876]: E1205 08:17:02.823434 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.856505 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.856593 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.856618 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.856642 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.856660 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.958808 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.958843 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.958851 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.958864 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:02 crc kubenswrapper[4876]: I1205 08:17:02.958873 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:02Z","lastTransitionTime":"2025-12-05T08:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.061886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.061973 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.061991 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.062014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.062034 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.164881 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.164965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.164985 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.165009 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.165027 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.268037 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.268099 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.268120 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.268149 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.268170 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.371101 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.371173 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.371192 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.371217 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.371235 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.474371 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.474416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.474426 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.474443 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.474454 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.576964 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.577011 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.577023 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.577040 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.577055 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.680290 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.680348 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.680365 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.680387 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.680404 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.784665 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.784754 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.784783 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.784820 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.784856 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.823456 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.823487 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.823596 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:03 crc kubenswrapper[4876]: E1205 08:17:03.823798 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:03 crc kubenswrapper[4876]: E1205 08:17:03.823997 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:03 crc kubenswrapper[4876]: E1205 08:17:03.824500 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.825300 4876 scope.go:117] "RemoveContainer" containerID="f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1" Dec 05 08:17:03 crc kubenswrapper[4876]: E1205 08:17:03.825892 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\"" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.844505 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.883098 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.888171 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.888238 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.888261 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.888291 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.888313 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.913714 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.928989 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.944300 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.956309 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.969591 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c5fa857-dcef-43e2-8f42-0b94fcd00541\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a056e141e0c391539ffd94a814372c68901b5e7ea03c7c696c333c5fa320142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.987990 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:03Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.990469 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.990528 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.990543 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.990564 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:03 crc kubenswrapper[4876]: I1205 08:17:03.990579 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:03Z","lastTransitionTime":"2025-12-05T08:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.003042 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.015802 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.028223 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.040230 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.052894 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.080370 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:50Z\\\",\\\"message\\\":\\\"rvices.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/community-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"d389393c-7ba9-422c-b3f5-06e391d537d2\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/community-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.189\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1205 08:16:50.071429 6853 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.093109 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.093172 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.093191 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.093215 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.093235 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.100431 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.121957 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:42Z\\\",\\\"message\\\":\\\"2025-12-05T08:15:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b\\\\n2025-12-05T08:15:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b to /host/opt/cni/bin/\\\\n2025-12-05T08:15:57Z [verbose] multus-daemon started\\\\n2025-12-05T08:15:57Z [verbose] Readiness Indicator file check\\\\n2025-12-05T08:16:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.144973 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.168632 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:04Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.196947 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.197012 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.197030 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.197053 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.197071 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.300621 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.300664 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.300674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.300687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.300697 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.403813 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.403887 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.403931 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.403956 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.403975 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.507106 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.507187 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.507210 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.507240 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.507265 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.609756 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.609813 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.609832 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.609855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.609873 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.712730 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.712792 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.712809 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.712835 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.712854 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.816087 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.816165 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.816183 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.816209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.816230 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.823295 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:04 crc kubenswrapper[4876]: E1205 08:17:04.823459 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.919848 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.919977 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.919997 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.920032 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:04 crc kubenswrapper[4876]: I1205 08:17:04.920058 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:04Z","lastTransitionTime":"2025-12-05T08:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.023385 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.023466 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.023486 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.023512 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.023536 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.127078 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.127445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.127492 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.127525 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.127552 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.230080 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.230157 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.230209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.230242 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.230268 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.333349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.333432 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.333458 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.333491 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.333515 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.435867 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.435971 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.435991 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.436014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.436031 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.538416 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.538498 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.538521 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.538544 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.538565 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.643604 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.643721 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.643741 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.643774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.643797 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.747095 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.747138 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.747149 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.747167 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.747177 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.823194 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.823292 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:05 crc kubenswrapper[4876]: E1205 08:17:05.823337 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.823463 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:05 crc kubenswrapper[4876]: E1205 08:17:05.823560 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:05 crc kubenswrapper[4876]: E1205 08:17:05.823705 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.841936 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tlgq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9911fb0-ecb1-429a-bf57-7c2ea8d2c505\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d600cd956febc8603d0e5876cb6c47be6b7599cba42f02c80f2de76155e3385\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnq84\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tlgq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.849387 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.849439 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.849454 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.849532 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.849550 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.855704 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7wg59" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79225fe3-5793-4316-92da-c227edb92d53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gqmth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7wg59\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.870047 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c5fa857-dcef-43e2-8f42-0b94fcd00541\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a056e141e0c391539ffd94a814372c68901b5e7ea03c7c696c333c5fa320142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://465a1cec9a9285aa448f6a23de86cc72af140e41eb98052345cbfb62fc857252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.887175 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.910168 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b9f0af3dc7701d6de0d9dd4f73a03a96194f45f60d45dc438967ca1fd631d7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f792665ef8fa44f8d2f5a9ca37877928a9d5223f44b190460301967ab539f25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.926093 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.939828 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n4dsp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae897c0d-b078-40d3-bbd4-aa2c279c89ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bacc08b62f0dea5e548f68850df96e3f060509f01bd25cfd97cd6352ae86c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fwhkw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n4dsp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.952734 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.952767 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.952781 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.952800 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.952813 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:05Z","lastTransitionTime":"2025-12-05T08:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.958420 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1229eb9b-4d2a-4301-bbdb-9b37e590ac11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68f9e16cb21fe5dc20a37c94219d4680ccbb21a4ff76300dc6a393ce574309b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6653ff451f56237ca92130fa52b7e0bbd9f3ad50fe32cba5f9adec3029b53a90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:16:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kl7xd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.974817 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4846bdfd-e3f9-43b1-bc83-b8d155ce793d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T08:15:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 08:15:48.480387 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 08:15:48.481516 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1496402630/tls.crt::/tmp/serving-cert-1496402630/tls.key\\\\\\\"\\\\nI1205 08:15:54.103213 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 08:15:54.108883 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 08:15:54.108953 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 08:15:54.108979 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 08:15:54.108985 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 08:15:54.128243 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 08:15:54.128263 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128267 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 08:15:54.128273 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 08:15:54.128276 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 08:15:54.128278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 08:15:54.128281 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 08:15:54.128420 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 08:15:54.132284 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:05 crc kubenswrapper[4876]: I1205 08:17:05.992366 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"238f3c94-018b-4993-8fb9-1508938f8f5c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f782d724653c66f5e20ebdf9cee5ec3392a353c2a295a28b880899be0e81bf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7751e730917ec320f67bc14f0f6c119cb8d88f8e72ab798f8eecc21cef157a66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc1247eeaf605d5cd02cf6ddd513155c6a6346e9e3ba89703edd57bd456783c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ce6c59aafa65b8ad250a420e30fc828d512aa4084f8bec86cf3a6aba32a5941\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:05Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.011859 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caab43d024e2fc64df90c30d55868345bf3c04f24dd8ae17e217015a587d7f89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.024625 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f6792109a3540992b9a28d21acc4199bdeaf7ed973525a2d8cb1550617599ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.039491 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42bd4d54-ee91-428e-bffa-1ac53c033ba9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5ae76179bdc15dbcbd365d0a3ff980ae881564c80582eda86ad7008e365e71b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16b8443bd714db858ffea892e48c1b3097d85f1ba66083ff9ad1c083fc6a1e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33f16aa2e2b00117fe917bb4ca7be3d5c88672aa0b33ec60805e32b389ca27c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.055327 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.056648 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.056682 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.056695 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.056715 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.056730 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.087124 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"446e4833-20af-47c0-b4e0-7116c3972366\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:50Z\\\",\\\"message\\\":\\\"rvices.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/community-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"d389393c-7ba9-422c-b3f5-06e391d537d2\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/community-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/community-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.189\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1205 08:16:50.071429 6853 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcg2f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-58npg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.108067 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-z44lm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"905a4c7a-1ee3-46ff-bd6b-04a34078015c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d503127d01905827eaf7148b4d6d1f84181a3eb302e01858768a1024342ffb2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0659c0fe7a2b226eae11b91c01f55d381aa6ef2c67e0adf82aca82c4728b72f5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfc4a8745c3e58e27b0d2a69a200284abea29f75cf877e2a8196408184523df9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cab4da934da7e257b380177e924c01ac3cb541d5847d96148716fabdd4192943\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c5629054e05cfabdf0de4587446b1014baad7d12fd420f90d17ab5ab1d8cfa7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:15:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e5fcd492ae3eccff691c922ccf72df10b02d4a4df12deeb9ba72b0ac56d747c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089b018095b2f56bb7135d11d522e3f27a47e3024b70db4c71cb99148b078aea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T08:16:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T08:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h79nh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-z44lm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.127694 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jwrfx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7274e0d-7f30-495c-800a-5dcbf1656b0d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T08:16:42Z\\\",\\\"message\\\":\\\"2025-12-05T08:15:56+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b\\\\n2025-12-05T08:15:56+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_034212c2-7978-4659-95fa-343f4907386b to /host/opt/cni/bin/\\\\n2025-12-05T08:15:57Z [verbose] multus-daemon started\\\\n2025-12-05T08:15:57Z [verbose] Readiness Indicator file check\\\\n2025-12-05T08:16:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98p7w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jwrfx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.144178 4876 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77322cc8-c6ab-4250-8098-9938309f0af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T08:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cca2649bfe5cdeeed154afc3f989fba6bcb49999f69fc485ee836f0d19c4a0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T08:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v9djw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T08:15:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xj8qx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:06Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.160285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.160366 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.160394 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.160426 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.160451 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.264383 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.264504 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.264570 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.264597 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.264624 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.367270 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.367331 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.367350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.367379 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.367399 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.470021 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.470093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.470118 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.470148 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.470171 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.572573 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.572633 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.572641 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.572655 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.572663 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.676114 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.676171 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.676187 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.676210 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.676227 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.778679 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.779105 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.779285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.779448 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.779623 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.823116 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:06 crc kubenswrapper[4876]: E1205 08:17:06.823317 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.883077 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.883168 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.883195 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.883227 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.883252 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.986261 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.986310 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.986327 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.986351 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:06 crc kubenswrapper[4876]: I1205 08:17:06.986368 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:06Z","lastTransitionTime":"2025-12-05T08:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.089571 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.089631 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.089649 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.089672 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.089689 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.192532 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.192590 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.192606 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.192634 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.192653 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.295848 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.295932 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.295952 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.295977 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.295994 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.399108 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.399162 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.399184 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.399213 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.399233 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.501853 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.501952 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.501975 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.502001 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.502020 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.604974 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.605087 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.605105 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.605178 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.605198 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.708347 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.708461 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.708483 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.708512 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.708534 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.812216 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.812280 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.812297 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.812324 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.812342 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.823565 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.823653 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.823596 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:07 crc kubenswrapper[4876]: E1205 08:17:07.823802 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:07 crc kubenswrapper[4876]: E1205 08:17:07.824005 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:07 crc kubenswrapper[4876]: E1205 08:17:07.824298 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.914765 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.914844 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.914861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.915016 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:07 crc kubenswrapper[4876]: I1205 08:17:07.915078 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:07Z","lastTransitionTime":"2025-12-05T08:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.017552 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.017595 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.017606 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.017622 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.017633 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.120696 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.120753 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.120763 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.120787 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.120801 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.224674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.224794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.224813 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.224841 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.224859 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.327222 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.327299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.327322 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.327356 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.327380 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.431048 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.431101 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.431110 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.431130 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.431141 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.534604 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.534687 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.534700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.534737 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.534752 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.638783 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.638825 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.638833 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.638849 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.638861 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.742445 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.742507 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.742526 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.742548 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.742566 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.823429 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:08 crc kubenswrapper[4876]: E1205 08:17:08.823963 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.846296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.846353 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.846369 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.846394 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.846410 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.949821 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.949918 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.949933 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.949963 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:08 crc kubenswrapper[4876]: I1205 08:17:08.949978 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:08Z","lastTransitionTime":"2025-12-05T08:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.009632 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.009686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.009703 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.009726 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.009744 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.030582 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.035804 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.035933 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.035965 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.035992 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.036015 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.055892 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.060864 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.060957 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.060984 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.061016 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.061037 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.084209 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.089580 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.089654 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.089673 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.089701 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.089722 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.111788 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.118027 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.118093 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.118110 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.118136 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.118155 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.140779 4876 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T08:17:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5aa7b9b5-1049-47fd-9576-7435395774d5\\\",\\\"systemUUID\\\":\\\"4a836e23-e8c7-4003-b121-650a726282d3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T08:17:09Z is after 2025-08-24T17:21:41Z" Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.141046 4876 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.143663 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.143715 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.143733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.143760 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.143781 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.247942 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.248009 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.248031 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.248063 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.248087 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.350696 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.350764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.350796 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.350835 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.350859 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.453453 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.453536 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.453570 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.453598 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.453617 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.556443 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.556514 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.556557 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.556590 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.556614 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.659816 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.659883 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.659936 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.659966 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.659985 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.762670 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.762733 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.762750 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.762774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.762791 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.823798 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.824086 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.824175 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.824573 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.824738 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:09 crc kubenswrapper[4876]: E1205 08:17:09.824884 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.865188 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.865254 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.865277 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.865303 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.865325 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.968024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.968100 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.968117 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.968147 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:09 crc kubenswrapper[4876]: I1205 08:17:09.968165 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:09Z","lastTransitionTime":"2025-12-05T08:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.071062 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.071150 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.071187 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.071218 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.071240 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.174351 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.174417 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.174434 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.174462 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.174479 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.277454 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.277536 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.277561 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.277591 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.277613 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.380432 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.380499 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.380515 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.380538 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.380555 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.483192 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.483229 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.483237 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.483250 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.483259 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.585937 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.585990 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.586003 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.586023 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.586035 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.689025 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.689068 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.689077 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.689090 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.689100 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.792341 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.792391 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.792407 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.792424 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.792436 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.823418 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:10 crc kubenswrapper[4876]: E1205 08:17:10.823597 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.895453 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.895515 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.895534 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.895560 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.895579 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.998885 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.998945 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.998954 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.998969 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:10 crc kubenswrapper[4876]: I1205 08:17:10.998979 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:10Z","lastTransitionTime":"2025-12-05T08:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.102537 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.102628 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.102647 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.102674 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.102693 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.205962 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.206033 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.206047 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.206067 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.206079 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.309073 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.309111 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.309141 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.309157 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.309170 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.412484 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.412603 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.412631 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.412702 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.412765 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.515460 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.515528 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.515550 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.515581 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.515602 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.619133 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.619196 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.619219 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.619250 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.619272 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.721349 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.721384 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.721395 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.721411 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.721422 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.823042 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.823126 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:11 crc kubenswrapper[4876]: E1205 08:17:11.823466 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.824092 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:11 crc kubenswrapper[4876]: E1205 08:17:11.824163 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.824240 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.824282 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: E1205 08:17:11.824281 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.824299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.824323 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.824346 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.839755 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.927101 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.927156 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.927179 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.927209 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:11 crc kubenswrapper[4876]: I1205 08:17:11.927236 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:11Z","lastTransitionTime":"2025-12-05T08:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.030430 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.030506 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.030523 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.030548 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.030565 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.134918 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.134970 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.134981 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.134996 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.135006 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.237344 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.237389 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.237401 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.237417 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.237429 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.324985 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:12 crc kubenswrapper[4876]: E1205 08:17:12.325170 4876 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:17:12 crc kubenswrapper[4876]: E1205 08:17:12.325298 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs podName:79225fe3-5793-4316-92da-c227edb92d53 nodeName:}" failed. No retries permitted until 2025-12-05 08:18:16.325274546 +0000 UTC m=+160.813939198 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs") pod "network-metrics-daemon-7wg59" (UID: "79225fe3-5793-4316-92da-c227edb92d53") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.339970 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.340037 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.340056 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.340080 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.340098 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.443114 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.443202 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.443224 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.443250 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.443277 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.546097 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.546161 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.546178 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.546205 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.546222 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.649135 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.649176 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.649185 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.649199 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.649208 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.751713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.751753 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.751766 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.751783 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.751795 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.822869 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:12 crc kubenswrapper[4876]: E1205 08:17:12.823020 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.855225 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.855259 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.855267 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.855280 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.855306 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.958664 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.958722 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.958745 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.958777 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:12 crc kubenswrapper[4876]: I1205 08:17:12.958799 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:12Z","lastTransitionTime":"2025-12-05T08:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.062286 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.062350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.062369 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.062397 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.062416 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.165935 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.165998 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.166015 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.166039 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.166060 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.268474 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.268541 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.268572 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.268598 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.268615 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.371218 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.371276 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.371294 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.371316 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.371333 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.474309 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.474370 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.474387 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.474410 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.474427 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.577740 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.577804 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.577822 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.577847 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.577866 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.681253 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.681336 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.681354 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.681383 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.681406 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.784589 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.784666 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.784714 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.784738 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.784752 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.823425 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.823435 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:13 crc kubenswrapper[4876]: E1205 08:17:13.823593 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.823619 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:13 crc kubenswrapper[4876]: E1205 08:17:13.823696 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:13 crc kubenswrapper[4876]: E1205 08:17:13.823940 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.892780 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.892872 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.892893 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.892960 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.892989 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.997234 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.997306 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.997323 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.997350 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:13 crc kubenswrapper[4876]: I1205 08:17:13.997368 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:13Z","lastTransitionTime":"2025-12-05T08:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.101794 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.101874 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.101940 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.101978 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.102003 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.205436 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.205509 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.205528 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.205553 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.205570 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.309601 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.309782 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.309810 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.309837 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.309855 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.413686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.413799 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.413819 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.413847 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.413867 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.532940 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.533012 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.533024 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.533041 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.533053 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.636609 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.636680 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.636699 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.636725 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.636744 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.739269 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.739319 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.739329 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.739348 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.739358 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.823415 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:14 crc kubenswrapper[4876]: E1205 08:17:14.823614 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.842178 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.842247 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.842269 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.842297 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.842320 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.944640 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.944694 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.944713 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.944738 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:14 crc kubenswrapper[4876]: I1205 08:17:14.944756 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:14Z","lastTransitionTime":"2025-12-05T08:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.054927 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.054986 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.055003 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.055026 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.055043 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.158351 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.158415 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.158432 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.158457 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.158478 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.262232 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.262288 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.262299 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.262318 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.262331 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.365793 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.365833 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.365843 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.365861 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.365873 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.468825 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.468867 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.468880 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.468916 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.468932 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.571414 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.571463 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.571476 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.571495 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.571508 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.675117 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.675220 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.675259 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.675313 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.675339 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.779150 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.779230 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.779255 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.779285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.779307 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.823633 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:15 crc kubenswrapper[4876]: E1205 08:17:15.823824 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.823965 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:15 crc kubenswrapper[4876]: E1205 08:17:15.824145 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.823973 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:15 crc kubenswrapper[4876]: E1205 08:17:15.824290 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.826586 4876 scope.go:117] "RemoveContainer" containerID="f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1" Dec 05 08:17:15 crc kubenswrapper[4876]: E1205 08:17:15.827094 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-58npg_openshift-ovn-kubernetes(446e4833-20af-47c0-b4e0-7116c3972366)\"" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.888369 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.888439 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.888464 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.888493 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.888515 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.922837 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.922807836 podStartE2EDuration="1m19.922807836s" podCreationTimestamp="2025-12-05 08:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:15.922468372 +0000 UTC m=+100.411133054" watchObservedRunningTime="2025-12-05 08:17:15.922807836 +0000 UTC m=+100.411472498" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.961623 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jwrfx" podStartSLOduration=81.961596839 podStartE2EDuration="1m21.961596839s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:15.945183706 +0000 UTC m=+100.433848338" watchObservedRunningTime="2025-12-05 08:17:15.961596839 +0000 UTC m=+100.450261471" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.986255 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-z44lm" podStartSLOduration=81.986232054 podStartE2EDuration="1m21.986232054s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:15.985826409 +0000 UTC m=+100.474491061" watchObservedRunningTime="2025-12-05 08:17:15.986232054 +0000 UTC m=+100.474896686" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.987163 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podStartSLOduration=81.987151534 podStartE2EDuration="1m21.987151534s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:15.959326084 +0000 UTC m=+100.447990726" watchObservedRunningTime="2025-12-05 08:17:15.987151534 +0000 UTC m=+100.475816176" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.995682 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.995751 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.995764 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.995781 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:15 crc kubenswrapper[4876]: I1205 08:17:15.995818 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:15Z","lastTransitionTime":"2025-12-05T08:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.042164 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-n4dsp" podStartSLOduration=82.042146328 podStartE2EDuration="1m22.042146328s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:16.0414707 +0000 UTC m=+100.530135362" watchObservedRunningTime="2025-12-05 08:17:16.042146328 +0000 UTC m=+100.530810950" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.062090 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tlgq9" podStartSLOduration=83.0620669 podStartE2EDuration="1m23.0620669s" podCreationTimestamp="2025-12-05 08:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:16.051966207 +0000 UTC m=+100.540630829" watchObservedRunningTime="2025-12-05 08:17:16.0620669 +0000 UTC m=+100.550731532" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.093350 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.093322439 podStartE2EDuration="1m22.093322439s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:16.092628731 +0000 UTC m=+100.581293393" watchObservedRunningTime="2025-12-05 08:17:16.093322439 +0000 UTC m=+100.581987081" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.094659 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=31.094649963 podStartE2EDuration="31.094649963s" podCreationTimestamp="2025-12-05 08:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:16.076050336 +0000 UTC m=+100.564714998" watchObservedRunningTime="2025-12-05 08:17:16.094649963 +0000 UTC m=+100.583314595" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.098345 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.098409 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.098426 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.098447 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.098464 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.107640 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=51.107615858 podStartE2EDuration="51.107615858s" podCreationTimestamp="2025-12-05 08:16:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:16.107408996 +0000 UTC m=+100.596073658" watchObservedRunningTime="2025-12-05 08:17:16.107615858 +0000 UTC m=+100.596280510" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.198537 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kl7xd" podStartSLOduration=82.198513162 podStartE2EDuration="1m22.198513162s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:16.198226469 +0000 UTC m=+100.686891101" watchObservedRunningTime="2025-12-05 08:17:16.198513162 +0000 UTC m=+100.687177824" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.200795 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.200838 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.200855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.200877 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.200942 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.303958 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.304034 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.304055 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.304501 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.304802 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.407886 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.407988 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.408010 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.408038 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.408055 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.510889 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.511009 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.511034 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.511065 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.511087 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.614600 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.614705 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.614724 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.614750 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.614799 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.718264 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.718307 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.718319 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.718336 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.718347 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.821625 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.821680 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.821696 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.821718 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.821734 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.823067 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:16 crc kubenswrapper[4876]: E1205 08:17:16.823392 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.925428 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.925600 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.925624 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.925650 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:16 crc kubenswrapper[4876]: I1205 08:17:16.925669 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:16Z","lastTransitionTime":"2025-12-05T08:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.028686 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.028756 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.028785 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.028816 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.028838 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.131987 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.132098 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.132128 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.132152 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.132169 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.236014 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.236077 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.236101 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.236132 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.236152 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.339582 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.339675 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.339696 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.339719 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.339738 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.442864 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.443062 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.443090 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.443121 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.443144 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.545741 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.545810 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.545829 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.545855 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.545872 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.648617 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.648672 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.648683 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.648700 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.648712 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.751637 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.751719 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.751743 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.751774 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.751797 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.823461 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.823493 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.824071 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:17 crc kubenswrapper[4876]: E1205 08:17:17.824264 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:17 crc kubenswrapper[4876]: E1205 08:17:17.824415 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:17 crc kubenswrapper[4876]: E1205 08:17:17.824602 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.854334 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.854425 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.854447 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.854472 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.854493 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.956882 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.956989 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.957027 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.957062 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:17 crc kubenswrapper[4876]: I1205 08:17:17.957088 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:17Z","lastTransitionTime":"2025-12-05T08:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.060529 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.060611 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.060634 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.060660 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.060678 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.163175 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.163285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.163307 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.163366 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.163387 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.266520 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.266598 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.266619 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.266651 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.266676 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.369022 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.369138 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.369167 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.369198 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.369221 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.472470 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.472551 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.472574 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.472609 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.472633 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.575760 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.575804 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.575813 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.575828 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.575837 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.678234 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.678285 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.678296 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.678313 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.678329 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.781057 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.781118 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.781137 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.781160 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.781177 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.823521 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:18 crc kubenswrapper[4876]: E1205 08:17:18.823646 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.883463 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.883547 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.883569 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.883595 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.883613 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.986851 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.986923 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.986936 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.986955 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:18 crc kubenswrapper[4876]: I1205 08:17:18.986992 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:18Z","lastTransitionTime":"2025-12-05T08:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.089460 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.089590 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.089607 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.089623 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.089634 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:19Z","lastTransitionTime":"2025-12-05T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.192376 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.192444 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.192461 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.192484 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.192501 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:19Z","lastTransitionTime":"2025-12-05T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.295684 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.295740 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.295758 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.295784 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.295802 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:19Z","lastTransitionTime":"2025-12-05T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.316522 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.316579 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.316603 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.316629 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.316649 4876 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T08:17:19Z","lastTransitionTime":"2025-12-05T08:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.384405 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=8.384377144 podStartE2EDuration="8.384377144s" podCreationTimestamp="2025-12-05 08:17:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:16.226147211 +0000 UTC m=+100.714811903" watchObservedRunningTime="2025-12-05 08:17:19.384377144 +0000 UTC m=+103.873041806" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.388471 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl"] Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.389076 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.391503 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.391624 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.392474 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.394033 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.531837 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db61de0a-8f07-4213-9d36-28f546d47284-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.531952 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/db61de0a-8f07-4213-9d36-28f546d47284-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.532209 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/db61de0a-8f07-4213-9d36-28f546d47284-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.532398 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db61de0a-8f07-4213-9d36-28f546d47284-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.532467 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/db61de0a-8f07-4213-9d36-28f546d47284-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.633633 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db61de0a-8f07-4213-9d36-28f546d47284-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.633683 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/db61de0a-8f07-4213-9d36-28f546d47284-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.633713 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db61de0a-8f07-4213-9d36-28f546d47284-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.633743 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/db61de0a-8f07-4213-9d36-28f546d47284-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.633794 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/db61de0a-8f07-4213-9d36-28f546d47284-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.634002 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/db61de0a-8f07-4213-9d36-28f546d47284-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.634049 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/db61de0a-8f07-4213-9d36-28f546d47284-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.636329 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/db61de0a-8f07-4213-9d36-28f546d47284-service-ca\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.640510 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db61de0a-8f07-4213-9d36-28f546d47284-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.658049 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db61de0a-8f07-4213-9d36-28f546d47284-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-p7pkl\" (UID: \"db61de0a-8f07-4213-9d36-28f546d47284\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.717674 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.823744 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.823813 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:19 crc kubenswrapper[4876]: E1205 08:17:19.824243 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:19 crc kubenswrapper[4876]: I1205 08:17:19.823842 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:19 crc kubenswrapper[4876]: E1205 08:17:19.824327 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:19 crc kubenswrapper[4876]: E1205 08:17:19.824620 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:20 crc kubenswrapper[4876]: I1205 08:17:20.367534 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" event={"ID":"db61de0a-8f07-4213-9d36-28f546d47284","Type":"ContainerStarted","Data":"516c6aa7ec63e75fcbc46eb3664614f44811d43192b8f55f9cdbfa079056db1f"} Dec 05 08:17:20 crc kubenswrapper[4876]: I1205 08:17:20.367607 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" event={"ID":"db61de0a-8f07-4213-9d36-28f546d47284","Type":"ContainerStarted","Data":"7e49b8f8008b2914217ae23ba7c689de10073ad95947dc5028d31a3579649b9a"} Dec 05 08:17:20 crc kubenswrapper[4876]: I1205 08:17:20.383385 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-p7pkl" podStartSLOduration=86.383366645 podStartE2EDuration="1m26.383366645s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:20.382311343 +0000 UTC m=+104.870975985" watchObservedRunningTime="2025-12-05 08:17:20.383366645 +0000 UTC m=+104.872031307" Dec 05 08:17:20 crc kubenswrapper[4876]: I1205 08:17:20.823754 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:20 crc kubenswrapper[4876]: E1205 08:17:20.824022 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:21 crc kubenswrapper[4876]: I1205 08:17:21.823124 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:21 crc kubenswrapper[4876]: I1205 08:17:21.823180 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:21 crc kubenswrapper[4876]: E1205 08:17:21.823310 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:21 crc kubenswrapper[4876]: I1205 08:17:21.823452 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:21 crc kubenswrapper[4876]: E1205 08:17:21.823581 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:21 crc kubenswrapper[4876]: E1205 08:17:21.823739 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:22 crc kubenswrapper[4876]: I1205 08:17:22.823206 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:22 crc kubenswrapper[4876]: E1205 08:17:22.823312 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:23 crc kubenswrapper[4876]: I1205 08:17:23.823214 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:23 crc kubenswrapper[4876]: I1205 08:17:23.823251 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:23 crc kubenswrapper[4876]: E1205 08:17:23.823360 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:23 crc kubenswrapper[4876]: I1205 08:17:23.823226 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:23 crc kubenswrapper[4876]: E1205 08:17:23.823477 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:23 crc kubenswrapper[4876]: E1205 08:17:23.823563 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:24 crc kubenswrapper[4876]: I1205 08:17:24.823300 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:24 crc kubenswrapper[4876]: E1205 08:17:24.823499 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:25 crc kubenswrapper[4876]: I1205 08:17:25.824435 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:25 crc kubenswrapper[4876]: E1205 08:17:25.824527 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:25 crc kubenswrapper[4876]: I1205 08:17:25.824631 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:25 crc kubenswrapper[4876]: I1205 08:17:25.824891 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:25 crc kubenswrapper[4876]: E1205 08:17:25.824988 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:25 crc kubenswrapper[4876]: E1205 08:17:25.825107 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:26 crc kubenswrapper[4876]: I1205 08:17:26.823223 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:26 crc kubenswrapper[4876]: E1205 08:17:26.823482 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:27 crc kubenswrapper[4876]: I1205 08:17:27.823147 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:27 crc kubenswrapper[4876]: I1205 08:17:27.823225 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:27 crc kubenswrapper[4876]: E1205 08:17:27.823385 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:27 crc kubenswrapper[4876]: I1205 08:17:27.823435 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:27 crc kubenswrapper[4876]: E1205 08:17:27.823569 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:27 crc kubenswrapper[4876]: E1205 08:17:27.823735 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:28 crc kubenswrapper[4876]: I1205 08:17:28.823197 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:28 crc kubenswrapper[4876]: E1205 08:17:28.823685 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.403118 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/1.log" Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.404025 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/0.log" Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.404107 4876 generic.go:334] "Generic (PLEG): container finished" podID="e7274e0d-7f30-495c-800a-5dcbf1656b0d" containerID="bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3" exitCode=1 Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.404150 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jwrfx" event={"ID":"e7274e0d-7f30-495c-800a-5dcbf1656b0d","Type":"ContainerDied","Data":"bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3"} Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.404196 4876 scope.go:117] "RemoveContainer" containerID="3c9653144f2b545e913c88a04f6150fa23b5b07278b447a08c1289ef9c44b0cc" Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.404739 4876 scope.go:117] "RemoveContainer" containerID="bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3" Dec 05 08:17:29 crc kubenswrapper[4876]: E1205 08:17:29.405008 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-jwrfx_openshift-multus(e7274e0d-7f30-495c-800a-5dcbf1656b0d)\"" pod="openshift-multus/multus-jwrfx" podUID="e7274e0d-7f30-495c-800a-5dcbf1656b0d" Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.823587 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:29 crc kubenswrapper[4876]: E1205 08:17:29.823706 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.823921 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:29 crc kubenswrapper[4876]: E1205 08:17:29.823991 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:29 crc kubenswrapper[4876]: I1205 08:17:29.824239 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:29 crc kubenswrapper[4876]: E1205 08:17:29.824306 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:30 crc kubenswrapper[4876]: I1205 08:17:30.411127 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/1.log" Dec 05 08:17:30 crc kubenswrapper[4876]: I1205 08:17:30.823705 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:30 crc kubenswrapper[4876]: E1205 08:17:30.823949 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:30 crc kubenswrapper[4876]: I1205 08:17:30.824927 4876 scope.go:117] "RemoveContainer" containerID="f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1" Dec 05 08:17:31 crc kubenswrapper[4876]: I1205 08:17:31.415852 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/3.log" Dec 05 08:17:31 crc kubenswrapper[4876]: I1205 08:17:31.418342 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerStarted","Data":"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9"} Dec 05 08:17:31 crc kubenswrapper[4876]: I1205 08:17:31.418671 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:17:31 crc kubenswrapper[4876]: I1205 08:17:31.779845 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podStartSLOduration=97.779826606 podStartE2EDuration="1m37.779826606s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:31.447800793 +0000 UTC m=+115.936465435" watchObservedRunningTime="2025-12-05 08:17:31.779826606 +0000 UTC m=+116.268491238" Dec 05 08:17:31 crc kubenswrapper[4876]: I1205 08:17:31.780137 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7wg59"] Dec 05 08:17:31 crc kubenswrapper[4876]: I1205 08:17:31.780232 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:31 crc kubenswrapper[4876]: E1205 08:17:31.780318 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:31 crc kubenswrapper[4876]: I1205 08:17:31.823407 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:31 crc kubenswrapper[4876]: I1205 08:17:31.823511 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:31 crc kubenswrapper[4876]: E1205 08:17:31.823553 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:31 crc kubenswrapper[4876]: E1205 08:17:31.823664 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:32 crc kubenswrapper[4876]: I1205 08:17:32.823110 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:32 crc kubenswrapper[4876]: E1205 08:17:32.823476 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:33 crc kubenswrapper[4876]: I1205 08:17:33.823382 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:33 crc kubenswrapper[4876]: I1205 08:17:33.823547 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:33 crc kubenswrapper[4876]: E1205 08:17:33.823716 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:33 crc kubenswrapper[4876]: I1205 08:17:33.823781 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:33 crc kubenswrapper[4876]: E1205 08:17:33.824031 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:33 crc kubenswrapper[4876]: E1205 08:17:33.824221 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:34 crc kubenswrapper[4876]: I1205 08:17:34.823957 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:34 crc kubenswrapper[4876]: E1205 08:17:34.824124 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:35 crc kubenswrapper[4876]: E1205 08:17:35.796960 4876 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 05 08:17:35 crc kubenswrapper[4876]: I1205 08:17:35.823569 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:35 crc kubenswrapper[4876]: I1205 08:17:35.823644 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:35 crc kubenswrapper[4876]: I1205 08:17:35.823725 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:35 crc kubenswrapper[4876]: E1205 08:17:35.826300 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:35 crc kubenswrapper[4876]: E1205 08:17:35.826431 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:35 crc kubenswrapper[4876]: E1205 08:17:35.826703 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:35 crc kubenswrapper[4876]: E1205 08:17:35.913986 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 08:17:36 crc kubenswrapper[4876]: I1205 08:17:36.823867 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:36 crc kubenswrapper[4876]: E1205 08:17:36.824103 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:37 crc kubenswrapper[4876]: I1205 08:17:37.822961 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:37 crc kubenswrapper[4876]: E1205 08:17:37.823113 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:37 crc kubenswrapper[4876]: I1205 08:17:37.823200 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:37 crc kubenswrapper[4876]: E1205 08:17:37.823469 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:37 crc kubenswrapper[4876]: I1205 08:17:37.824024 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:37 crc kubenswrapper[4876]: E1205 08:17:37.824355 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:38 crc kubenswrapper[4876]: I1205 08:17:38.823554 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:38 crc kubenswrapper[4876]: E1205 08:17:38.823756 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:39 crc kubenswrapper[4876]: I1205 08:17:39.823176 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:39 crc kubenswrapper[4876]: I1205 08:17:39.823271 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:39 crc kubenswrapper[4876]: I1205 08:17:39.823369 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:39 crc kubenswrapper[4876]: E1205 08:17:39.823361 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:39 crc kubenswrapper[4876]: E1205 08:17:39.823466 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:39 crc kubenswrapper[4876]: E1205 08:17:39.823680 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:40 crc kubenswrapper[4876]: I1205 08:17:40.823677 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:40 crc kubenswrapper[4876]: E1205 08:17:40.824327 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:40 crc kubenswrapper[4876]: I1205 08:17:40.825811 4876 scope.go:117] "RemoveContainer" containerID="bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3" Dec 05 08:17:40 crc kubenswrapper[4876]: E1205 08:17:40.915103 4876 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 08:17:41 crc kubenswrapper[4876]: I1205 08:17:41.453873 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/1.log" Dec 05 08:17:41 crc kubenswrapper[4876]: I1205 08:17:41.454071 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jwrfx" event={"ID":"e7274e0d-7f30-495c-800a-5dcbf1656b0d","Type":"ContainerStarted","Data":"341c70583ca8b1edc982af298aae9fa74350b607faec9525aa3fe1a89c48e098"} Dec 05 08:17:41 crc kubenswrapper[4876]: I1205 08:17:41.822931 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:41 crc kubenswrapper[4876]: I1205 08:17:41.822979 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:41 crc kubenswrapper[4876]: I1205 08:17:41.823049 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:41 crc kubenswrapper[4876]: E1205 08:17:41.823181 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:41 crc kubenswrapper[4876]: E1205 08:17:41.823360 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:41 crc kubenswrapper[4876]: E1205 08:17:41.823532 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:42 crc kubenswrapper[4876]: I1205 08:17:42.823638 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:42 crc kubenswrapper[4876]: E1205 08:17:42.823994 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:43 crc kubenswrapper[4876]: I1205 08:17:43.823802 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:43 crc kubenswrapper[4876]: I1205 08:17:43.823867 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:43 crc kubenswrapper[4876]: E1205 08:17:43.823976 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:43 crc kubenswrapper[4876]: I1205 08:17:43.824153 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:43 crc kubenswrapper[4876]: E1205 08:17:43.824245 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:43 crc kubenswrapper[4876]: E1205 08:17:43.824274 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:44 crc kubenswrapper[4876]: I1205 08:17:44.823052 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:44 crc kubenswrapper[4876]: E1205 08:17:44.823277 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 08:17:45 crc kubenswrapper[4876]: I1205 08:17:45.823513 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:45 crc kubenswrapper[4876]: I1205 08:17:45.823609 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:45 crc kubenswrapper[4876]: I1205 08:17:45.824709 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:45 crc kubenswrapper[4876]: E1205 08:17:45.824691 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 08:17:45 crc kubenswrapper[4876]: E1205 08:17:45.824879 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7wg59" podUID="79225fe3-5793-4316-92da-c227edb92d53" Dec 05 08:17:45 crc kubenswrapper[4876]: E1205 08:17:45.825050 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 08:17:46 crc kubenswrapper[4876]: I1205 08:17:46.823792 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:17:46 crc kubenswrapper[4876]: I1205 08:17:46.827932 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 08:17:46 crc kubenswrapper[4876]: I1205 08:17:46.828420 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 08:17:47 crc kubenswrapper[4876]: I1205 08:17:47.823360 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:17:47 crc kubenswrapper[4876]: I1205 08:17:47.823409 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:17:47 crc kubenswrapper[4876]: I1205 08:17:47.823424 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:17:47 crc kubenswrapper[4876]: I1205 08:17:47.825707 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 08:17:47 crc kubenswrapper[4876]: I1205 08:17:47.826239 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 08:17:47 crc kubenswrapper[4876]: I1205 08:17:47.826299 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 08:17:47 crc kubenswrapper[4876]: I1205 08:17:47.826840 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.069551 4876 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.121021 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.121662 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6nzhg"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.121893 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.123180 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.126185 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.127022 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.128521 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.133405 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.134031 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.134309 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.134496 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.135067 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.135689 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.136231 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.136558 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.137883 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.138264 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.138680 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.139998 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.140334 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.141650 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vqw7m"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.142325 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w85pf"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.143046 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4cv77"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.143124 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.143299 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vqw7m" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.146164 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.150090 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.150866 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.152567 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.153412 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.161569 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-r8p7s"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.161737 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.162079 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.174306 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.174969 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.175069 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.175260 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.175888 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-plzs7"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.176490 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.178315 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.178536 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.178775 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.178995 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.179394 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.179610 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.179782 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.180131 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.180197 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.180338 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.180652 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.180810 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.181029 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.181174 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.182167 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.186934 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.187092 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.187182 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zc7rw"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.187765 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.188603 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.188765 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.188822 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.197564 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.197915 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.198869 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.199194 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.200105 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.204229 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.204293 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.205608 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.205770 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.205981 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-client-ca\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.208001 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/859647d8-54ea-4892-ae54-2179375e1ae0-serving-cert\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.208330 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.225090 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226232 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-images\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226371 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-config\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226779 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62c18aef-7ad2-4616-beb0-d286f175f1d1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226818 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r49sz\" (UniqueName: \"kubernetes.io/projected/62c18aef-7ad2-4616-beb0-d286f175f1d1-kube-api-access-r49sz\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226846 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t75m\" (UniqueName: \"kubernetes.io/projected/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-kube-api-access-5t75m\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226867 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-config\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226937 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15da8e1-2f29-41fa-8dae-dce111e40262-serving-cert\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226962 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.226982 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227006 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c18aef-7ad2-4616-beb0-d286f175f1d1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227046 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-config\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227067 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9qng\" (UniqueName: \"kubernetes.io/projected/859647d8-54ea-4892-ae54-2179375e1ae0-kube-api-access-x9qng\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227109 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bqpm\" (UniqueName: \"kubernetes.io/projected/b15da8e1-2f29-41fa-8dae-dce111e40262-kube-api-access-4bqpm\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227132 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s7z9\" (UniqueName: \"kubernetes.io/projected/86dbc349-8d0e-4524-a7a6-0292689ad46c-kube-api-access-2s7z9\") pod \"dns-operator-744455d44c-4cv77\" (UID: \"86dbc349-8d0e-4524-a7a6-0292689ad46c\") " pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227154 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-client-ca\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227175 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrrx5\" (UniqueName: \"kubernetes.io/projected/5d3b7e9c-38f3-435a-a765-c9c2f3b360e8-kube-api-access-jrrx5\") pod \"downloads-7954f5f757-vqw7m\" (UID: \"5d3b7e9c-38f3-435a-a765-c9c2f3b360e8\") " pod="openshift-console/downloads-7954f5f757-vqw7m" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227206 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86dbc349-8d0e-4524-a7a6-0292689ad46c-metrics-tls\") pod \"dns-operator-744455d44c-4cv77\" (UID: \"86dbc349-8d0e-4524-a7a6-0292689ad46c\") " pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227226 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ggxp\" (UniqueName: \"kubernetes.io/projected/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-kube-api-access-6ggxp\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227602 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.227795 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.228430 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.229025 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.231141 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.231203 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.231422 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.231706 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.231714 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.232399 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.235582 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.243028 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.243927 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.244085 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.244140 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.244285 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.244386 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.245117 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.245207 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.245280 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.245344 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.245400 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.245435 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.247975 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.247998 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.248044 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.248054 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.249202 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.248106 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.248608 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.249395 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pqx48"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.248649 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.248691 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.251791 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.252604 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.253108 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.253390 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.253652 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.256268 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.257038 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.258808 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.259594 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.260679 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.261302 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.262531 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.262737 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.275170 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vzhwk"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.282139 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.283437 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.283708 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.279369 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.284522 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.292513 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.292643 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.292810 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.293951 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.294671 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.294785 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.294839 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.294997 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.295088 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.295274 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.296122 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.296263 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.296418 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.296565 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.296648 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.296772 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.297106 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.297180 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.297393 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.297451 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.297521 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.297780 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.298047 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.298361 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.299265 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.299567 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-6rtbf"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.300108 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8czl9"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.300542 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sklcd"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.301034 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.301364 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.301887 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.302396 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.303797 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.304229 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gj226"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.304710 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.305063 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.305158 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.306461 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.307575 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.313587 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6nzhg"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.313873 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.314492 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.315010 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.315348 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.315705 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.316350 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-scljl"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.318021 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.318069 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.321620 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z6kjz"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.321791 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.322099 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-96pbm"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.322208 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.322842 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.323046 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.323267 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mm2m6"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.323436 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.323675 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.324016 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.324061 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.324164 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vqw7m"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.325151 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.324203 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.325788 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.325815 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.325828 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.325841 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w85pf"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.325857 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-plzs7"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.325927 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327683 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/81bf7719-efce-4d93-8e77-6273ff226a5e-auth-proxy-config\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327710 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bhgfr\" (UID: \"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327730 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ggxp\" (UniqueName: \"kubernetes.io/projected/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-kube-api-access-6ggxp\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327751 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-service-ca\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327769 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-etcd-client\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327811 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327830 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86dbc349-8d0e-4524-a7a6-0292689ad46c-metrics-tls\") pod \"dns-operator-744455d44c-4cv77\" (UID: \"86dbc349-8d0e-4524-a7a6-0292689ad46c\") " pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327844 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tlxz\" (UniqueName: \"kubernetes.io/projected/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-kube-api-access-7tlxz\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327862 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327876 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92e719e9-c64e-4e54-a002-427ebe228761-audit-dir\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327892 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnv4v\" (UniqueName: \"kubernetes.io/projected/92e719e9-c64e-4e54-a002-427ebe228761-kube-api-access-jnv4v\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327920 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327934 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327950 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed2fd37f-5667-4509-a422-92dba00024c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327965 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327979 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.327993 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-serving-cert\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329133 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329176 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.328007 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329494 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329521 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-serving-cert\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329538 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv9p2\" (UniqueName: \"kubernetes.io/projected/81bf7719-efce-4d93-8e77-6273ff226a5e-kube-api-access-hv9p2\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329552 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-etcd-client\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329567 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-audit-dir\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329580 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329598 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-config\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329614 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329634 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7r5t\" (UniqueName: \"kubernetes.io/projected/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-kube-api-access-b7r5t\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329650 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329677 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329706 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqfjz\" (UniqueName: \"kubernetes.io/projected/e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160-kube-api-access-vqfjz\") pod \"cluster-samples-operator-665b6dd947-bhgfr\" (UID: \"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329730 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-audit-policies\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329752 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-policies\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329772 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329797 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329816 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-encryption-config\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329840 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-client-ca\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329860 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/859647d8-54ea-4892-ae54-2179375e1ae0-serving-cert\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329880 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-config\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329915 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329934 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-images\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329948 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329962 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329980 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-config\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.329995 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-etcd-serving-ca\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330011 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r49sz\" (UniqueName: \"kubernetes.io/projected/62c18aef-7ad2-4616-beb0-d286f175f1d1-kube-api-access-r49sz\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330107 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-oauth-config\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330125 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-config\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330412 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330140 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62c18aef-7ad2-4616-beb0-d286f175f1d1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330450 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330471 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t75m\" (UniqueName: \"kubernetes.io/projected/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-kube-api-access-5t75m\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330487 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-config\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330506 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330534 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-audit\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330567 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15da8e1-2f29-41fa-8dae-dce111e40262-serving-cert\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330588 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330604 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330622 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330639 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330656 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c18aef-7ad2-4616-beb0-d286f175f1d1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330693 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-service-ca-bundle\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330710 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/81bf7719-efce-4d93-8e77-6273ff226a5e-machine-approver-tls\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330725 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-oauth-serving-cert\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330738 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/92e719e9-c64e-4e54-a002-427ebe228761-node-pullsecrets\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330753 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330769 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-image-import-ca\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330783 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330797 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-serving-cert\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330811 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-trusted-ca-bundle\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330824 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-dir\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330847 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bf7719-efce-4d93-8e77-6273ff226a5e-config\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330862 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmv6c\" (UniqueName: \"kubernetes.io/projected/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-kube-api-access-vmv6c\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330888 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-config\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330925 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9qng\" (UniqueName: \"kubernetes.io/projected/859647d8-54ea-4892-ae54-2179375e1ae0-kube-api-access-x9qng\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330945 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed2fd37f-5667-4509-a422-92dba00024c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330962 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkbqn\" (UniqueName: \"kubernetes.io/projected/ed2fd37f-5667-4509-a422-92dba00024c7-kube-api-access-gkbqn\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330978 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2qwt\" (UniqueName: \"kubernetes.io/projected/646287d9-6646-42a6-a5e2-9100f5b9b680-kube-api-access-s2qwt\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.330993 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed2fd37f-5667-4509-a422-92dba00024c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331007 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331023 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2qtd\" (UniqueName: \"kubernetes.io/projected/4c5a2919-0f9c-4f28-b6e4-ab288f991483-kube-api-access-p2qtd\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331047 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-client-ca\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bqpm\" (UniqueName: \"kubernetes.io/projected/b15da8e1-2f29-41fa-8dae-dce111e40262-kube-api-access-4bqpm\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331163 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s7z9\" (UniqueName: \"kubernetes.io/projected/86dbc349-8d0e-4524-a7a6-0292689ad46c-kube-api-access-2s7z9\") pod \"dns-operator-744455d44c-4cv77\" (UID: \"86dbc349-8d0e-4524-a7a6-0292689ad46c\") " pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331179 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/646287d9-6646-42a6-a5e2-9100f5b9b680-serving-cert\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331194 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331211 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrrx5\" (UniqueName: \"kubernetes.io/projected/5d3b7e9c-38f3-435a-a765-c9c2f3b360e8-kube-api-access-jrrx5\") pod \"downloads-7954f5f757-vqw7m\" (UID: \"5d3b7e9c-38f3-435a-a765-c9c2f3b360e8\") " pod="openshift-console/downloads-7954f5f757-vqw7m" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331228 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrzmz\" (UniqueName: \"kubernetes.io/projected/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-kube-api-access-wrzmz\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331242 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-encryption-config\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331925 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4cv77"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.331940 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-client-ca\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.332369 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-config\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.334408 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86dbc349-8d0e-4524-a7a6-0292689ad46c-metrics-tls\") pod \"dns-operator-744455d44c-4cv77\" (UID: \"86dbc349-8d0e-4524-a7a6-0292689ad46c\") " pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.334681 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/859647d8-54ea-4892-ae54-2179375e1ae0-serving-cert\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.335287 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.335341 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-images\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.347460 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c18aef-7ad2-4616-beb0-d286f175f1d1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.350625 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-config\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.351658 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15da8e1-2f29-41fa-8dae-dce111e40262-serving-cert\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.352188 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.353215 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.353883 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-r8p7s"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.354008 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vzhwk"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.354700 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-client-ca\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.357056 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-config\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.357379 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62c18aef-7ad2-4616-beb0-d286f175f1d1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.365537 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.369260 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.369772 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pqx48"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.371686 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.372956 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.375965 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-tx9gn"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.376651 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tx9gn" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.377273 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.378220 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.379416 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z6kjz"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.380450 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gj226"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.381486 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.383678 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.383718 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.384399 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.384583 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.385461 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zc7rw"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.386390 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.387686 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8czl9"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.388658 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-96pbm"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.389617 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.390578 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.391576 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rd9lr"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.392395 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.392908 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-r7wwl"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.393769 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.393971 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-scljl"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.395037 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.396300 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mm2m6"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.398667 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.399939 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tx9gn"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.400825 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sklcd"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.401846 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.404154 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.404379 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rd9lr"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.405609 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xb5cw"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.406884 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xb5cw"] Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.407016 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.424076 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432338 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-etcd-client\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432373 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432406 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-service-ca\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432427 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92e719e9-c64e-4e54-a002-427ebe228761-audit-dir\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432448 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tlxz\" (UniqueName: \"kubernetes.io/projected/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-kube-api-access-7tlxz\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432469 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432491 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432515 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnv4v\" (UniqueName: \"kubernetes.io/projected/92e719e9-c64e-4e54-a002-427ebe228761-kube-api-access-jnv4v\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432537 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432558 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed2fd37f-5667-4509-a422-92dba00024c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432580 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432605 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ba874267-c442-44a8-9ed6-c8e0b1bc7505-srv-cert\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432627 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5742635-f4b3-47ee-88be-d59720792b25-secret-volume\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432666 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432687 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-serving-cert\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432711 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9986f6ed-67ea-4ac8-bae1-0e782bd83112-config\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.432746 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92e719e9-c64e-4e54-a002-427ebe228761-audit-dir\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.433815 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g29l9\" (UniqueName: \"kubernetes.io/projected/f0573958-5bcc-4deb-b42b-e2d65ea550b5-kube-api-access-g29l9\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.433851 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-serving-cert\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.433942 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdhzr\" (UniqueName: \"kubernetes.io/projected/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-kube-api-access-wdhzr\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.434082 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.434104 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.434116 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-service-ca\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.434165 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.434279 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.434590 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed2fd37f-5667-4509-a422-92dba00024c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435226 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435232 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-etcd-client\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.434191 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435315 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-etcd-client\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435340 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-audit-dir\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435364 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435401 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c16d20e-2b84-4436-8c6d-1bf4cea72706-profile-collector-cert\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435414 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-audit-dir\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435427 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv9p2\" (UniqueName: \"kubernetes.io/projected/81bf7719-efce-4d93-8e77-6273ff226a5e-kube-api-access-hv9p2\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435450 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2htcn\" (UniqueName: \"kubernetes.io/projected/808f78c5-f306-4f4a-a823-7e114bb9be9f-kube-api-access-2htcn\") pod \"migrator-59844c95c7-gfdrk\" (UID: \"808f78c5-f306-4f4a-a823-7e114bb9be9f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435478 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-config\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435501 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435524 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7r5t\" (UniqueName: \"kubernetes.io/projected/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-kube-api-access-b7r5t\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435546 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.435568 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436082 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-serving-cert\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436215 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-config\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436316 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqfjz\" (UniqueName: \"kubernetes.io/projected/e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160-kube-api-access-vqfjz\") pod \"cluster-samples-operator-665b6dd947-bhgfr\" (UID: \"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436362 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-audit-policies\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436318 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436385 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-policies\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436408 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436447 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436473 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-encryption-config\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436496 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-config\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436521 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436544 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zxbc\" (UniqueName: \"kubernetes.io/projected/36087227-ee02-4dd0-b801-2ced90e4173f-kube-api-access-5zxbc\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436569 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436593 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436616 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436642 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-config\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436665 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-etcd-serving-ca\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436689 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436712 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-config\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436746 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-oauth-config\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436772 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436792 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-audit-policies\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436801 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436828 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436851 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436867 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436892 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-audit\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436947 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436972 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzxbp\" (UniqueName: \"kubernetes.io/projected/2c16d20e-2b84-4436-8c6d-1bf4cea72706-kube-api-access-wzxbp\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.436999 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437023 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-service-ca-bundle\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437048 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/81bf7719-efce-4d93-8e77-6273ff226a5e-machine-approver-tls\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437068 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-oauth-serving-cert\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437090 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/92e719e9-c64e-4e54-a002-427ebe228761-node-pullsecrets\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437112 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-image-import-ca\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437141 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437172 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437203 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dwxb\" (UniqueName: \"kubernetes.io/projected/9986f6ed-67ea-4ac8-bae1-0e782bd83112-kube-api-access-9dwxb\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437222 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-policies\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437230 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-trusted-ca-bundle\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437254 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-dir\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437278 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf5fb\" (UniqueName: \"kubernetes.io/projected/b5742635-f4b3-47ee-88be-d59720792b25-kube-api-access-lf5fb\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437450 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-serving-cert\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437471 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bf7719-efce-4d93-8e77-6273ff226a5e-config\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437494 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmv6c\" (UniqueName: \"kubernetes.io/projected/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-kube-api-access-vmv6c\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437515 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c16d20e-2b84-4436-8c6d-1bf4cea72706-srv-cert\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437555 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed2fd37f-5667-4509-a422-92dba00024c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437588 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkbqn\" (UniqueName: \"kubernetes.io/projected/ed2fd37f-5667-4509-a422-92dba00024c7-kube-api-access-gkbqn\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437610 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2qwt\" (UniqueName: \"kubernetes.io/projected/646287d9-6646-42a6-a5e2-9100f5b9b680-kube-api-access-s2qwt\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437632 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437652 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-cabundle\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437675 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed2fd37f-5667-4509-a422-92dba00024c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437698 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437720 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2qtd\" (UniqueName: \"kubernetes.io/projected/4c5a2919-0f9c-4f28-b6e4-ab288f991483-kube-api-access-p2qtd\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437741 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-key\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437780 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/646287d9-6646-42a6-a5e2-9100f5b9b680-serving-cert\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437801 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437831 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-config\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438144 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.437833 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ba874267-c442-44a8-9ed6-c8e0b1bc7505-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438214 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz98q\" (UniqueName: \"kubernetes.io/projected/ba874267-c442-44a8-9ed6-c8e0b1bc7505-kube-api-access-nz98q\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438255 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrzmz\" (UniqueName: \"kubernetes.io/projected/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-kube-api-access-wrzmz\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438264 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-dir\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438285 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-encryption-config\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438311 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/81bf7719-efce-4d93-8e77-6273ff226a5e-auth-proxy-config\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438326 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-serving-cert\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438338 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bhgfr\" (UID: \"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438363 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9986f6ed-67ea-4ac8-bae1-0e782bd83112-serving-cert\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438370 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-etcd-client\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438620 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438722 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.438915 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-etcd-serving-ca\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.439142 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/81bf7719-efce-4d93-8e77-6273ff226a5e-auth-proxy-config\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.439265 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.439339 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-config\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.439625 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/92e719e9-c64e-4e54-a002-427ebe228761-encryption-config\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.440200 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.440259 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-audit\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.440376 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-image-import-ca\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.440617 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/92e719e9-c64e-4e54-a002-427ebe228761-node-pullsecrets\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.440750 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.440883 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-oauth-serving-cert\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.441181 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.441289 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81bf7719-efce-4d93-8e77-6273ff226a5e-config\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.441670 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.441698 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/646287d9-6646-42a6-a5e2-9100f5b9b680-service-ca-bundle\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.441747 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-trusted-ca-bundle\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.441889 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92e719e9-c64e-4e54-a002-427ebe228761-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.443113 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.443279 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.443288 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.443409 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed2fd37f-5667-4509-a422-92dba00024c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.443661 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-encryption-config\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.444096 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bhgfr\" (UID: \"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.444879 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.445165 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-serving-cert\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.445245 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-oauth-config\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.445304 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.445587 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.445871 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/81bf7719-efce-4d93-8e77-6273ff226a5e-machine-approver-tls\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.446045 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.446151 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/646287d9-6646-42a6-a5e2-9100f5b9b680-serving-cert\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.464971 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.484336 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.504396 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.524468 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543488 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543534 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543556 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-config\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543631 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543653 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543692 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzxbp\" (UniqueName: \"kubernetes.io/projected/2c16d20e-2b84-4436-8c6d-1bf4cea72706-kube-api-access-wzxbp\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543718 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dwxb\" (UniqueName: \"kubernetes.io/projected/9986f6ed-67ea-4ac8-bae1-0e782bd83112-kube-api-access-9dwxb\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543738 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf5fb\" (UniqueName: \"kubernetes.io/projected/b5742635-f4b3-47ee-88be-d59720792b25-kube-api-access-lf5fb\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543768 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c16d20e-2b84-4436-8c6d-1bf4cea72706-srv-cert\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543828 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543851 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-cabundle\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543887 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-key\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543959 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ba874267-c442-44a8-9ed6-c8e0b1bc7505-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.543979 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz98q\" (UniqueName: \"kubernetes.io/projected/ba874267-c442-44a8-9ed6-c8e0b1bc7505-kube-api-access-nz98q\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544011 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9986f6ed-67ea-4ac8-bae1-0e782bd83112-serving-cert\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544063 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ba874267-c442-44a8-9ed6-c8e0b1bc7505-srv-cert\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544087 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5742635-f4b3-47ee-88be-d59720792b25-secret-volume\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544107 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9986f6ed-67ea-4ac8-bae1-0e782bd83112-config\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544127 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g29l9\" (UniqueName: \"kubernetes.io/projected/f0573958-5bcc-4deb-b42b-e2d65ea550b5-kube-api-access-g29l9\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544146 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdhzr\" (UniqueName: \"kubernetes.io/projected/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-kube-api-access-wdhzr\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544171 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c16d20e-2b84-4436-8c6d-1bf4cea72706-profile-collector-cert\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544206 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2htcn\" (UniqueName: \"kubernetes.io/projected/808f78c5-f306-4f4a-a823-7e114bb9be9f-kube-api-access-2htcn\") pod \"migrator-59844c95c7-gfdrk\" (UID: \"808f78c5-f306-4f4a-a823-7e114bb9be9f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544263 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544313 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.544361 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zxbc\" (UniqueName: \"kubernetes.io/projected/36087227-ee02-4dd0-b801-2ced90e4173f-kube-api-access-5zxbc\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.564540 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.584235 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.603841 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.624604 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.643686 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.664191 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.683668 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.704154 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.708957 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.724258 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.735359 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-config\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.744199 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.764536 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.784402 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.804560 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.824312 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.845274 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.864290 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.884715 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.904625 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.925177 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.944576 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.966533 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 08:17:50 crc kubenswrapper[4876]: I1205 08:17:50.985602 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.004852 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.025146 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.045559 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.064763 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.092815 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.104486 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.124889 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.145539 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.165097 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.184639 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.205114 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.225154 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.245000 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.259539 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.276113 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.285608 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.287378 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.305513 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.322990 4876 request.go:700] Waited for 1.007408068s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serving-cert&limit=500&resourceVersion=0 Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.324785 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.341345 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ba874267-c442-44a8-9ed6-c8e0b1bc7505-srv-cert\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.345289 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.365391 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.385972 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.405632 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.419651 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5742635-f4b3-47ee-88be-d59720792b25-secret-volume\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.419712 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2c16d20e-2b84-4436-8c6d-1bf4cea72706-profile-collector-cert\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.421161 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ba874267-c442-44a8-9ed6-c8e0b1bc7505-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.426158 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.445926 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.465424 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.479889 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2c16d20e-2b84-4436-8c6d-1bf4cea72706-srv-cert\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.485351 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.505286 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.524667 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544076 4876 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544165 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume podName:b5742635-f4b3-47ee-88be-d59720792b25 nodeName:}" failed. No retries permitted until 2025-12-05 08:17:52.044141331 +0000 UTC m=+136.532805963 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume") pod "collect-profiles-29415375-796gt" (UID: "b5742635-f4b3-47ee-88be-d59720792b25") : failed to sync configmap cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544443 4876 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544503 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-cabundle podName:f0573958-5bcc-4deb-b42b-e2d65ea550b5 nodeName:}" failed. No retries permitted until 2025-12-05 08:17:52.044486899 +0000 UTC m=+136.533151531 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-cabundle") pod "service-ca-9c57cc56f-z6kjz" (UID: "f0573958-5bcc-4deb-b42b-e2d65ea550b5") : failed to sync configmap cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544542 4876 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544576 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-config podName:8eaba618-88b4-4d8d-9b87-fdabef6b7aad nodeName:}" failed. No retries permitted until 2025-12-05 08:17:52.044565281 +0000 UTC m=+136.533229913 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-config") pod "kube-storage-version-migrator-operator-b67b599dd-tbspv" (UID: "8eaba618-88b4-4d8d-9b87-fdabef6b7aad") : failed to sync configmap cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544616 4876 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544654 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9986f6ed-67ea-4ac8-bae1-0e782bd83112-serving-cert podName:9986f6ed-67ea-4ac8-bae1-0e782bd83112 nodeName:}" failed. No retries permitted until 2025-12-05 08:17:52.044643933 +0000 UTC m=+136.533308565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/9986f6ed-67ea-4ac8-bae1-0e782bd83112-serving-cert") pod "service-ca-operator-777779d784-96pbm" (UID: "9986f6ed-67ea-4ac8-bae1-0e782bd83112") : failed to sync secret cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544683 4876 secret.go:188] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544717 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-serving-cert podName:8eaba618-88b4-4d8d-9b87-fdabef6b7aad nodeName:}" failed. No retries permitted until 2025-12-05 08:17:52.044707295 +0000 UTC m=+136.533371937 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-serving-cert") pod "kube-storage-version-migrator-operator-b67b599dd-tbspv" (UID: "8eaba618-88b4-4d8d-9b87-fdabef6b7aad") : failed to sync secret cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544738 4876 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544774 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-key podName:f0573958-5bcc-4deb-b42b-e2d65ea550b5 nodeName:}" failed. No retries permitted until 2025-12-05 08:17:52.044762816 +0000 UTC m=+136.533427448 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-key") pod "service-ca-9c57cc56f-z6kjz" (UID: "f0573958-5bcc-4deb-b42b-e2d65ea550b5") : failed to sync secret cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544811 4876 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: E1205 08:17:51.544842 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9986f6ed-67ea-4ac8-bae1-0e782bd83112-config podName:9986f6ed-67ea-4ac8-bae1-0e782bd83112 nodeName:}" failed. No retries permitted until 2025-12-05 08:17:52.044832458 +0000 UTC m=+136.533497090 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/9986f6ed-67ea-4ac8-bae1-0e782bd83112-config") pod "service-ca-operator-777779d784-96pbm" (UID: "9986f6ed-67ea-4ac8-bae1-0e782bd83112") : failed to sync configmap cache: timed out waiting for the condition Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.545049 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.565380 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.585140 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.605348 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.624800 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.645735 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.665546 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.685887 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.704697 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.725341 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.746061 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.765334 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.784399 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.805983 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.825572 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.845368 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.865018 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.885456 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.905352 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.925060 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.972818 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ggxp\" (UniqueName: \"kubernetes.io/projected/dbb7d698-fe25-4fe1-8207-67bd7af52f7d-kube-api-access-6ggxp\") pod \"openshift-config-operator-7777fb866f-h9vc6\" (UID: \"dbb7d698-fe25-4fe1-8207-67bd7af52f7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:51 crc kubenswrapper[4876]: I1205 08:17:51.992328 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t75m\" (UniqueName: \"kubernetes.io/projected/dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf-kube-api-access-5t75m\") pod \"machine-api-operator-5694c8668f-w85pf\" (UID: \"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.012963 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r49sz\" (UniqueName: \"kubernetes.io/projected/62c18aef-7ad2-4616-beb0-d286f175f1d1-kube-api-access-r49sz\") pod \"openshift-apiserver-operator-796bbdcf4f-7nmrr\" (UID: \"62c18aef-7ad2-4616-beb0-d286f175f1d1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.027129 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.035446 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9qng\" (UniqueName: \"kubernetes.io/projected/859647d8-54ea-4892-ae54-2179375e1ae0-kube-api-access-x9qng\") pod \"route-controller-manager-6576b87f9c-2c65z\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.055796 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s7z9\" (UniqueName: \"kubernetes.io/projected/86dbc349-8d0e-4524-a7a6-0292689ad46c-kube-api-access-2s7z9\") pod \"dns-operator-744455d44c-4cv77\" (UID: \"86dbc349-8d0e-4524-a7a6-0292689ad46c\") " pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.068688 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-cabundle\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.068775 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-key\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.068892 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9986f6ed-67ea-4ac8-bae1-0e782bd83112-serving-cert\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.069048 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9986f6ed-67ea-4ac8-bae1-0e782bd83112-config\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.069281 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.069395 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.069495 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.071343 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-cabundle\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.071543 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.071657 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9986f6ed-67ea-4ac8-bae1-0e782bd83112-config\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.071997 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.074540 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bqpm\" (UniqueName: \"kubernetes.io/projected/b15da8e1-2f29-41fa-8dae-dce111e40262-kube-api-access-4bqpm\") pod \"controller-manager-879f6c89f-6nzhg\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.074689 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.075280 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9986f6ed-67ea-4ac8-bae1-0e782bd83112-serving-cert\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.075797 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f0573958-5bcc-4deb-b42b-e2d65ea550b5-signing-key\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.089972 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrrx5\" (UniqueName: \"kubernetes.io/projected/5d3b7e9c-38f3-435a-a765-c9c2f3b360e8-kube-api-access-jrrx5\") pod \"downloads-7954f5f757-vqw7m\" (UID: \"5d3b7e9c-38f3-435a-a765-c9c2f3b360e8\") " pod="openshift-console/downloads-7954f5f757-vqw7m" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.099324 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vqw7m" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.107183 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.112439 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.125155 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.146062 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.165956 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.186357 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.204791 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.226111 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.245500 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.256821 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.266224 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.270872 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-w85pf"] Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.278217 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:52 crc kubenswrapper[4876]: W1205 08:17:52.279542 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbb73726_8f4a_4d39_b22f_2eb5c3dd5bdf.slice/crio-260e545deec9fcf48d2ad31122bff65e2d4890b1fa5e45141bd045ef5bb53113 WatchSource:0}: Error finding container 260e545deec9fcf48d2ad31122bff65e2d4890b1fa5e45141bd045ef5bb53113: Status 404 returned error can't find the container with id 260e545deec9fcf48d2ad31122bff65e2d4890b1fa5e45141bd045ef5bb53113 Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.285664 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.288720 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.307153 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.307488 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vqw7m"] Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.323112 4876 request.go:700] Waited for 1.915829591s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/secrets?fieldSelector=metadata.name%3Dcsi-hostpath-provisioner-sa-dockercfg-qd74k&limit=500&resourceVersion=0 Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.324656 4876 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.334263 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6"] Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.344645 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.349122 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.381104 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.401584 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tlxz\" (UniqueName: \"kubernetes.io/projected/438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f-kube-api-access-7tlxz\") pod \"cluster-image-registry-operator-dc59b4c8b-qbjhf\" (UID: \"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.418507 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnv4v\" (UniqueName: \"kubernetes.io/projected/92e719e9-c64e-4e54-a002-427ebe228761-kube-api-access-jnv4v\") pod \"apiserver-76f77b778f-zc7rw\" (UID: \"92e719e9-c64e-4e54-a002-427ebe228761\") " pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.442593 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7r5t\" (UniqueName: \"kubernetes.io/projected/a29c5fbc-f737-48f5-87c8-7c7bebc8648d-kube-api-access-b7r5t\") pod \"apiserver-7bbb656c7d-lw86s\" (UID: \"a29c5fbc-f737-48f5-87c8-7c7bebc8648d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.458880 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c4ffecba-7d0e-42ca-8b18-9792a7ef34a4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lrlj\" (UID: \"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.472951 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr"] Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.478331 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv9p2\" (UniqueName: \"kubernetes.io/projected/81bf7719-efce-4d93-8e77-6273ff226a5e-kube-api-access-hv9p2\") pod \"machine-approver-56656f9798-c25rq\" (UID: \"81bf7719-efce-4d93-8e77-6273ff226a5e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.480371 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.486360 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.504730 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqfjz\" (UniqueName: \"kubernetes.io/projected/e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160-kube-api-access-vqfjz\") pod \"cluster-samples-operator-665b6dd947-bhgfr\" (UID: \"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.506453 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.508657 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" event={"ID":"dbb7d698-fe25-4fe1-8207-67bd7af52f7d","Type":"ContainerStarted","Data":"23d0ad328bda39ebd8f1798e4b06f86e502060aa62196fab195818b7dc7772b9"} Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.512652 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6nzhg"] Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.512752 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vqw7m" event={"ID":"5d3b7e9c-38f3-435a-a765-c9c2f3b360e8","Type":"ContainerStarted","Data":"d45a47dc67f64f9d3ba33af50f83c13f2a81998aef16ebefb9e247ddef7e6959"} Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.512930 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.515508 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" event={"ID":"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf","Type":"ContainerStarted","Data":"260e545deec9fcf48d2ad31122bff65e2d4890b1fa5e45141bd045ef5bb53113"} Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.517261 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2qtd\" (UniqueName: \"kubernetes.io/projected/4c5a2919-0f9c-4f28-b6e4-ab288f991483-kube-api-access-p2qtd\") pod \"oauth-openshift-558db77b4-r8p7s\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.519949 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.526096 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.540281 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2qwt\" (UniqueName: \"kubernetes.io/projected/646287d9-6646-42a6-a5e2-9100f5b9b680-kube-api-access-s2qwt\") pod \"authentication-operator-69f744f599-pqx48\" (UID: \"646287d9-6646-42a6-a5e2-9100f5b9b680\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.543016 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z"] Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.560610 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrzmz\" (UniqueName: \"kubernetes.io/projected/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-kube-api-access-wrzmz\") pod \"console-f9d7485db-plzs7\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.577935 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkbqn\" (UniqueName: \"kubernetes.io/projected/ed2fd37f-5667-4509-a422-92dba00024c7-kube-api-access-gkbqn\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.582938 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4cv77"] Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.596219 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed2fd37f-5667-4509-a422-92dba00024c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g29vs\" (UID: \"ed2fd37f-5667-4509-a422-92dba00024c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.617182 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmv6c\" (UniqueName: \"kubernetes.io/projected/43d50654-e5bb-495a-8c3e-89bbb7dc3bf4-kube-api-access-vmv6c\") pod \"openshift-controller-manager-operator-756b6f6bc6-pb7jc\" (UID: \"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.642123 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d4482e5-4766-4e46-a4a4-f02af8e4ac1c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-cmsv8\" (UID: \"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.658087 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf5fb\" (UniqueName: \"kubernetes.io/projected/b5742635-f4b3-47ee-88be-d59720792b25-kube-api-access-lf5fb\") pod \"collect-profiles-29415375-796gt\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.678472 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzxbp\" (UniqueName: \"kubernetes.io/projected/2c16d20e-2b84-4436-8c6d-1bf4cea72706-kube-api-access-wzxbp\") pod \"catalog-operator-68c6474976-4dcfx\" (UID: \"2c16d20e-2b84-4436-8c6d-1bf4cea72706\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.698541 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dwxb\" (UniqueName: \"kubernetes.io/projected/9986f6ed-67ea-4ac8-bae1-0e782bd83112-kube-api-access-9dwxb\") pod \"service-ca-operator-777779d784-96pbm\" (UID: \"9986f6ed-67ea-4ac8-bae1-0e782bd83112\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.719495 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz98q\" (UniqueName: \"kubernetes.io/projected/ba874267-c442-44a8-9ed6-c8e0b1bc7505-kube-api-access-nz98q\") pod \"olm-operator-6b444d44fb-sch8c\" (UID: \"ba874267-c442-44a8-9ed6-c8e0b1bc7505\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.733062 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.733168 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.759934 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.760976 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdhzr\" (UniqueName: \"kubernetes.io/projected/8eaba618-88b4-4d8d-9b87-fdabef6b7aad-kube-api-access-wdhzr\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbspv\" (UID: \"8eaba618-88b4-4d8d-9b87-fdabef6b7aad\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.762324 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g29l9\" (UniqueName: \"kubernetes.io/projected/f0573958-5bcc-4deb-b42b-e2d65ea550b5-kube-api-access-g29l9\") pod \"service-ca-9c57cc56f-z6kjz\" (UID: \"f0573958-5bcc-4deb-b42b-e2d65ea550b5\") " pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.780814 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zxbc\" (UniqueName: \"kubernetes.io/projected/36087227-ee02-4dd0-b801-2ced90e4173f-kube-api-access-5zxbc\") pod \"marketplace-operator-79b997595-gj226\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.795731 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.798989 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" Dec 05 08:17:52 crc kubenswrapper[4876]: I1205 08:17:52.801258 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2htcn\" (UniqueName: \"kubernetes.io/projected/808f78c5-f306-4f4a-a823-7e114bb9be9f-kube-api-access-2htcn\") pod \"migrator-59844c95c7-gfdrk\" (UID: \"808f78c5-f306-4f4a-a823-7e114bb9be9f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.016042 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.016181 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.016319 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.016707 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.017608 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.018169 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.019063 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.019645 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" Dec 05 08:17:53 crc kubenswrapper[4876]: W1205 08:17:53.032512 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb15da8e1_2f29_41fa_8dae_dce111e40262.slice/crio-7b55d99eb74589b2033f67cd7863aec6c07c315cdf1ccb3cca565a50c41abf1c WatchSource:0}: Error finding container 7b55d99eb74589b2033f67cd7863aec6c07c315cdf1ccb3cca565a50c41abf1c: Status 404 returned error can't find the container with id 7b55d99eb74589b2033f67cd7863aec6c07c315cdf1ccb3cca565a50c41abf1c Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.032744 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.035700 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/06abb292-ed23-4de1-9974-c487e7bb2933-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfjn\" (UID: \"06abb292-ed23-4de1-9974-c487e7bb2933\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036069 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-service-ca\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036152 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cdca41c-e056-44f6-9f06-6c129d1368cd-proxy-tls\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036197 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6a5866e-5650-4e91-9003-3d7537713d05-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mm2m6\" (UID: \"c6a5866e-5650-4e91-9003-3d7537713d05\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036263 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c9db69d-c1c4-4326-abb1-16d92bd6441d-trusted-ca\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036327 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-registry-certificates\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036368 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmw76\" (UniqueName: \"kubernetes.io/projected/6b11e75d-3346-4570-8266-59223cbb0619-kube-api-access-pmw76\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036413 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhm65\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-kube-api-access-xhm65\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036454 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhw8k\" (UniqueName: \"kubernetes.io/projected/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-kube-api-access-nhw8k\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036495 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-stats-auth\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036533 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cdca41c-e056-44f6-9f06-6c129d1368cd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036578 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-ca\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036642 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xqs7\" (UniqueName: \"kubernetes.io/projected/fed5f23f-513b-449f-adaa-d65ce9beb793-kube-api-access-2xqs7\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036700 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-default-certificate\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036770 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w57h7\" (UniqueName: \"kubernetes.io/projected/06abb292-ed23-4de1-9974-c487e7bb2933-kube-api-access-w57h7\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfjn\" (UID: \"06abb292-ed23-4de1-9974-c487e7bb2933\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.036925 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/907ad872-3805-403d-9649-11e7090bb789-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.037103 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b11e75d-3346-4570-8266-59223cbb0619-webhook-cert\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.037371 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s2bf\" (UniqueName: \"kubernetes.io/projected/9e452b45-5f78-4f52-b96e-83aa3dee3376-kube-api-access-4s2bf\") pod \"package-server-manager-789f6589d5-tvfkn\" (UID: \"9e452b45-5f78-4f52-b96e-83aa3dee3376\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.037468 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07af6910-37e6-401a-a863-c3de22d0e39f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.037513 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-config\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.038778 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d8130b4-8e6a-462c-bc65-d58567c00cd1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.050759 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07af6910-37e6-401a-a863-c3de22d0e39f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.050875 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d8130b4-8e6a-462c-bc65-d58567c00cd1-proxy-tls\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.050967 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-bound-sa-token\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.051023 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4rqt\" (UniqueName: \"kubernetes.io/projected/c6a5866e-5650-4e91-9003-3d7537713d05-kube-api-access-h4rqt\") pod \"multus-admission-controller-857f4d67dd-mm2m6\" (UID: \"c6a5866e-5650-4e91-9003-3d7537713d05\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.051128 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/907ad872-3805-403d-9649-11e7090bb789-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.051304 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-client\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.051432 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krfbf\" (UniqueName: \"kubernetes.io/projected/4d8130b4-8e6a-462c-bc65-d58567c00cd1-kube-api-access-krfbf\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.051604 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpmqh\" (UniqueName: \"kubernetes.io/projected/5cdca41c-e056-44f6-9f06-6c129d1368cd-kube-api-access-xpmqh\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.051678 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4d8130b4-8e6a-462c-bc65-d58567c00cd1-images\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.051746 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.051793 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6b11e75d-3346-4570-8266-59223cbb0619-tmpfs\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.052173 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-serving-cert\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.052235 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e452b45-5f78-4f52-b96e-83aa3dee3376-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-tvfkn\" (UID: \"9e452b45-5f78-4f52-b96e-83aa3dee3376\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.052261 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-registry-tls\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.052287 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx5ls\" (UniqueName: \"kubernetes.io/projected/6c9db69d-c1c4-4326-abb1-16d92bd6441d-kube-api-access-nx5ls\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.052414 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9db69d-c1c4-4326-abb1-16d92bd6441d-config\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.052586 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:53.552560276 +0000 UTC m=+138.041224938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.053706 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c9db69d-c1c4-4326-abb1-16d92bd6441d-serving-cert\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.055186 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-trusted-ca\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.056074 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07af6910-37e6-401a-a863-c3de22d0e39f-config\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.056154 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b11e75d-3346-4570-8266-59223cbb0619-apiservice-cert\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.056188 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fed5f23f-513b-449f-adaa-d65ce9beb793-service-ca-bundle\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.056218 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-metrics-certs\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.157279 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158041 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xqs7\" (UniqueName: \"kubernetes.io/projected/fed5f23f-513b-449f-adaa-d65ce9beb793-kube-api-access-2xqs7\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158076 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-default-certificate\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158104 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w57h7\" (UniqueName: \"kubernetes.io/projected/06abb292-ed23-4de1-9974-c487e7bb2933-kube-api-access-w57h7\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfjn\" (UID: \"06abb292-ed23-4de1-9974-c487e7bb2933\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158140 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-csi-data-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158202 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d2073f7-877e-4ba2-bc9b-a8d34d15a447-cert\") pod \"ingress-canary-tx9gn\" (UID: \"1d2073f7-877e-4ba2-bc9b-a8d34d15a447\") " pod="openshift-ingress-canary/ingress-canary-tx9gn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158258 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/907ad872-3805-403d-9649-11e7090bb789-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158286 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b11e75d-3346-4570-8266-59223cbb0619-webhook-cert\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158325 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s2bf\" (UniqueName: \"kubernetes.io/projected/9e452b45-5f78-4f52-b96e-83aa3dee3376-kube-api-access-4s2bf\") pod \"package-server-manager-789f6589d5-tvfkn\" (UID: \"9e452b45-5f78-4f52-b96e-83aa3dee3376\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158375 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07af6910-37e6-401a-a863-c3de22d0e39f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158402 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-config\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158442 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6qk6\" (UniqueName: \"kubernetes.io/projected/56f14ba6-fcaf-45fa-84d8-9a4461dda473-kube-api-access-c6qk6\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158479 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d8130b4-8e6a-462c-bc65-d58567c00cd1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158507 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-plugins-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158562 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07af6910-37e6-401a-a863-c3de22d0e39f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.158606 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:53.658577627 +0000 UTC m=+138.147242279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158659 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d8130b4-8e6a-462c-bc65-d58567c00cd1-proxy-tls\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158725 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-bound-sa-token\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158766 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4rqt\" (UniqueName: \"kubernetes.io/projected/c6a5866e-5650-4e91-9003-3d7537713d05-kube-api-access-h4rqt\") pod \"multus-admission-controller-857f4d67dd-mm2m6\" (UID: \"c6a5866e-5650-4e91-9003-3d7537713d05\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158823 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/56f14ba6-fcaf-45fa-84d8-9a4461dda473-node-bootstrap-token\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.158895 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/907ad872-3805-403d-9649-11e7090bb789-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159005 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-client\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159069 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krfbf\" (UniqueName: \"kubernetes.io/projected/4d8130b4-8e6a-462c-bc65-d58567c00cd1-kube-api-access-krfbf\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159112 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwnzn\" (UniqueName: \"kubernetes.io/projected/eece9379-11d6-44ff-9ca9-11ef43acad82-kube-api-access-wwnzn\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159163 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpmqh\" (UniqueName: \"kubernetes.io/projected/5cdca41c-e056-44f6-9f06-6c129d1368cd-kube-api-access-xpmqh\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159192 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4d8130b4-8e6a-462c-bc65-d58567c00cd1-images\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159225 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-serving-cert\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159267 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159303 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6b11e75d-3346-4570-8266-59223cbb0619-tmpfs\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159338 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e452b45-5f78-4f52-b96e-83aa3dee3376-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-tvfkn\" (UID: \"9e452b45-5f78-4f52-b96e-83aa3dee3376\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159373 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-registry-tls\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159403 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx5ls\" (UniqueName: \"kubernetes.io/projected/6c9db69d-c1c4-4326-abb1-16d92bd6441d-kube-api-access-nx5ls\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159470 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9db69d-c1c4-4326-abb1-16d92bd6441d-config\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159539 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c9db69d-c1c4-4326-abb1-16d92bd6441d-serving-cert\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159591 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fnjx\" (UniqueName: \"kubernetes.io/projected/1d2073f7-877e-4ba2-bc9b-a8d34d15a447-kube-api-access-4fnjx\") pod \"ingress-canary-tx9gn\" (UID: \"1d2073f7-877e-4ba2-bc9b-a8d34d15a447\") " pod="openshift-ingress-canary/ingress-canary-tx9gn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159655 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eece9379-11d6-44ff-9ca9-11ef43acad82-config-volume\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159707 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-trusted-ca\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159755 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-mountpoint-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159800 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07af6910-37e6-401a-a863-c3de22d0e39f-config\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159822 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/56f14ba6-fcaf-45fa-84d8-9a4461dda473-certs\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159849 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b11e75d-3346-4570-8266-59223cbb0619-apiservice-cert\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159874 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fed5f23f-513b-449f-adaa-d65ce9beb793-service-ca-bundle\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.159919 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-metrics-certs\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160003 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eece9379-11d6-44ff-9ca9-11ef43acad82-metrics-tls\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160027 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/06abb292-ed23-4de1-9974-c487e7bb2933-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfjn\" (UID: \"06abb292-ed23-4de1-9974-c487e7bb2933\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160048 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-service-ca\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160075 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cdca41c-e056-44f6-9f06-6c129d1368cd-proxy-tls\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160155 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6a5866e-5650-4e91-9003-3d7537713d05-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mm2m6\" (UID: \"c6a5866e-5650-4e91-9003-3d7537713d05\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160180 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c9db69d-c1c4-4326-abb1-16d92bd6441d-trusted-ca\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160206 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-socket-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160257 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-registry-certificates\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160285 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmw76\" (UniqueName: \"kubernetes.io/projected/6b11e75d-3346-4570-8266-59223cbb0619-kube-api-access-pmw76\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160340 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhm65\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-kube-api-access-xhm65\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160386 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhw8k\" (UniqueName: \"kubernetes.io/projected/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-kube-api-access-nhw8k\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160417 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-stats-auth\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160678 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwmgp\" (UniqueName: \"kubernetes.io/projected/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-kube-api-access-xwmgp\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160714 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cdca41c-e056-44f6-9f06-6c129d1368cd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160765 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-ca\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.160798 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-registration-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.161572 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/907ad872-3805-403d-9649-11e7090bb789-ca-trust-extracted\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.165236 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4d8130b4-8e6a-462c-bc65-d58567c00cd1-images\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.166493 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d8130b4-8e6a-462c-bc65-d58567c00cd1-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.168527 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:53.668505556 +0000 UTC m=+138.157170208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.168749 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-registry-certificates\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.169065 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07af6910-37e6-401a-a863-c3de22d0e39f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.169930 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07af6910-37e6-401a-a863-c3de22d0e39f-config\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.170148 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-metrics-certs\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.170813 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-default-certificate\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.170958 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6b11e75d-3346-4570-8266-59223cbb0619-webhook-cert\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.171576 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fed5f23f-513b-449f-adaa-d65ce9beb793-service-ca-bundle\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.173236 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6b11e75d-3346-4570-8266-59223cbb0619-tmpfs\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.174686 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c9db69d-c1c4-4326-abb1-16d92bd6441d-trusted-ca\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.175041 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e452b45-5f78-4f52-b96e-83aa3dee3376-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-tvfkn\" (UID: \"9e452b45-5f78-4f52-b96e-83aa3dee3376\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.175422 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6b11e75d-3346-4570-8266-59223cbb0619-apiservice-cert\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.175782 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9db69d-c1c4-4326-abb1-16d92bd6441d-config\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.176328 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/907ad872-3805-403d-9649-11e7090bb789-installation-pull-secrets\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.176333 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-trusted-ca\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.176865 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c6a5866e-5650-4e91-9003-3d7537713d05-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mm2m6\" (UID: \"c6a5866e-5650-4e91-9003-3d7537713d05\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.177102 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d8130b4-8e6a-462c-bc65-d58567c00cd1-proxy-tls\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.177685 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cdca41c-e056-44f6-9f06-6c129d1368cd-proxy-tls\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.178026 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/fed5f23f-513b-449f-adaa-d65ce9beb793-stats-auth\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.178487 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-registry-tls\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.178669 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cdca41c-e056-44f6-9f06-6c129d1368cd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.179026 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-client\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.179289 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-ca\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.179531 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-etcd-service-ca\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.179701 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-config\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.179832 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/06abb292-ed23-4de1-9974-c487e7bb2933-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfjn\" (UID: \"06abb292-ed23-4de1-9974-c487e7bb2933\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.180533 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c9db69d-c1c4-4326-abb1-16d92bd6441d-serving-cert\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.184128 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-serving-cert\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.185455 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s2bf\" (UniqueName: \"kubernetes.io/projected/9e452b45-5f78-4f52-b96e-83aa3dee3376-kube-api-access-4s2bf\") pod \"package-server-manager-789f6589d5-tvfkn\" (UID: \"9e452b45-5f78-4f52-b96e-83aa3dee3376\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.204809 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w57h7\" (UniqueName: \"kubernetes.io/projected/06abb292-ed23-4de1-9974-c487e7bb2933-kube-api-access-w57h7\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfjn\" (UID: \"06abb292-ed23-4de1-9974-c487e7bb2933\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.221570 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xqs7\" (UniqueName: \"kubernetes.io/projected/fed5f23f-513b-449f-adaa-d65ce9beb793-kube-api-access-2xqs7\") pod \"router-default-5444994796-6rtbf\" (UID: \"fed5f23f-513b-449f-adaa-d65ce9beb793\") " pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.228219 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.249654 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krfbf\" (UniqueName: \"kubernetes.io/projected/4d8130b4-8e6a-462c-bc65-d58567c00cd1-kube-api-access-krfbf\") pod \"machine-config-operator-74547568cd-5jxj6\" (UID: \"4d8130b4-8e6a-462c-bc65-d58567c00cd1\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.262387 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.263062 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:53.763020299 +0000 UTC m=+138.251684961 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.263841 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.263939 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fnjx\" (UniqueName: \"kubernetes.io/projected/1d2073f7-877e-4ba2-bc9b-a8d34d15a447-kube-api-access-4fnjx\") pod \"ingress-canary-tx9gn\" (UID: \"1d2073f7-877e-4ba2-bc9b-a8d34d15a447\") " pod="openshift-ingress-canary/ingress-canary-tx9gn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.263959 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eece9379-11d6-44ff-9ca9-11ef43acad82-config-volume\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264001 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-mountpoint-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264024 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/56f14ba6-fcaf-45fa-84d8-9a4461dda473-certs\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264047 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eece9379-11d6-44ff-9ca9-11ef43acad82-metrics-tls\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264091 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-socket-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264129 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwmgp\" (UniqueName: \"kubernetes.io/projected/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-kube-api-access-xwmgp\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264171 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-registration-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264203 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-csi-data-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264248 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d2073f7-877e-4ba2-bc9b-a8d34d15a447-cert\") pod \"ingress-canary-tx9gn\" (UID: \"1d2073f7-877e-4ba2-bc9b-a8d34d15a447\") " pod="openshift-ingress-canary/ingress-canary-tx9gn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264284 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6qk6\" (UniqueName: \"kubernetes.io/projected/56f14ba6-fcaf-45fa-84d8-9a4461dda473-kube-api-access-c6qk6\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264301 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-plugins-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264326 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-mountpoint-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264405 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/56f14ba6-fcaf-45fa-84d8-9a4461dda473-node-bootstrap-token\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264439 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwnzn\" (UniqueName: \"kubernetes.io/projected/eece9379-11d6-44ff-9ca9-11ef43acad82-kube-api-access-wwnzn\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264550 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-registration-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264753 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-csi-data-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264798 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-socket-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.264941 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-plugins-dir\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.265564 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:53.765549882 +0000 UTC m=+138.254214504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.266236 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eece9379-11d6-44ff-9ca9-11ef43acad82-config-volume\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.272453 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.273955 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/56f14ba6-fcaf-45fa-84d8-9a4461dda473-certs\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.275690 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d2073f7-877e-4ba2-bc9b-a8d34d15a447-cert\") pod \"ingress-canary-tx9gn\" (UID: \"1d2073f7-877e-4ba2-bc9b-a8d34d15a447\") " pod="openshift-ingress-canary/ingress-canary-tx9gn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.278949 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/eece9379-11d6-44ff-9ca9-11ef43acad82-metrics-tls\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.280077 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4rqt\" (UniqueName: \"kubernetes.io/projected/c6a5866e-5650-4e91-9003-3d7537713d05-kube-api-access-h4rqt\") pod \"multus-admission-controller-857f4d67dd-mm2m6\" (UID: \"c6a5866e-5650-4e91-9003-3d7537713d05\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.286288 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-bound-sa-token\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.298477 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/56f14ba6-fcaf-45fa-84d8-9a4461dda473-node-bootstrap-token\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.299601 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpmqh\" (UniqueName: \"kubernetes.io/projected/5cdca41c-e056-44f6-9f06-6c129d1368cd-kube-api-access-xpmqh\") pod \"machine-config-controller-84d6567774-scljl\" (UID: \"5cdca41c-e056-44f6-9f06-6c129d1368cd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.305706 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.314097 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.327881 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmw76\" (UniqueName: \"kubernetes.io/projected/6b11e75d-3346-4570-8266-59223cbb0619-kube-api-access-pmw76\") pod \"packageserver-d55dfcdfc-nhprd\" (UID: \"6b11e75d-3346-4570-8266-59223cbb0619\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.344444 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhm65\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-kube-api-access-xhm65\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.365167 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhw8k\" (UniqueName: \"kubernetes.io/projected/d726e1ea-2af3-4ade-80cb-e92fe8bd16d8-kube-api-access-nhw8k\") pod \"etcd-operator-b45778765-vzhwk\" (UID: \"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.365579 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.365963 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:53.865948542 +0000 UTC m=+138.354613164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.385829 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx5ls\" (UniqueName: \"kubernetes.io/projected/6c9db69d-c1c4-4326-abb1-16d92bd6441d-kube-api-access-nx5ls\") pod \"console-operator-58897d9998-8czl9\" (UID: \"6c9db69d-c1c4-4326-abb1-16d92bd6441d\") " pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.404064 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07af6910-37e6-401a-a863-c3de22d0e39f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-l827b\" (UID: \"07af6910-37e6-401a-a863-c3de22d0e39f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.442156 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fnjx\" (UniqueName: \"kubernetes.io/projected/1d2073f7-877e-4ba2-bc9b-a8d34d15a447-kube-api-access-4fnjx\") pod \"ingress-canary-tx9gn\" (UID: \"1d2073f7-877e-4ba2-bc9b-a8d34d15a447\") " pod="openshift-ingress-canary/ingress-canary-tx9gn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.466526 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.467615 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:53.967591874 +0000 UTC m=+138.456256586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.474702 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.480934 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.484232 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwnzn\" (UniqueName: \"kubernetes.io/projected/eece9379-11d6-44ff-9ca9-11ef43acad82-kube-api-access-wwnzn\") pod \"dns-default-rd9lr\" (UID: \"eece9379-11d6-44ff-9ca9-11ef43acad82\") " pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.485215 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwmgp\" (UniqueName: \"kubernetes.io/projected/86e64b2f-3a7e-4221-8f80-6423ccd3e1dd-kube-api-access-xwmgp\") pod \"csi-hostpathplugin-xb5cw\" (UID: \"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd\") " pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.503579 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.512513 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.519086 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vqw7m" event={"ID":"5d3b7e9c-38f3-435a-a765-c9c2f3b360e8","Type":"ContainerStarted","Data":"28f650f5f6138f42216b1dda545e73da672b49c6f4ecae18bcbffbb5c8d854ff"} Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.519094 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6qk6\" (UniqueName: \"kubernetes.io/projected/56f14ba6-fcaf-45fa-84d8-9a4461dda473-kube-api-access-c6qk6\") pod \"machine-config-server-r7wwl\" (UID: \"56f14ba6-fcaf-45fa-84d8-9a4461dda473\") " pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.519338 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vqw7m" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.520073 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" event={"ID":"859647d8-54ea-4892-ae54-2179375e1ae0","Type":"ContainerStarted","Data":"8112206fa85e10508f0990c604e387f6c984eb21f9b7d8e9442fdbdaa7ea7090"} Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.523119 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-vqw7m container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.523159 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vqw7m" podUID="5d3b7e9c-38f3-435a-a765-c9c2f3b360e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.523532 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" event={"ID":"62c18aef-7ad2-4616-beb0-d286f175f1d1","Type":"ContainerStarted","Data":"568a87c1f8f52b36641f2eac1e4565c21b1bb36c4330fdd046a27c7cbad7a30e"} Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.527083 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" event={"ID":"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf","Type":"ContainerStarted","Data":"b5c530ca3264ff1e280312054208d890631d231f1206f08ed0875012279bb535"} Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.529486 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" event={"ID":"81bf7719-efce-4d93-8e77-6273ff226a5e","Type":"ContainerStarted","Data":"546ab35842337684e421e7acf760c514bd45e9fc0eeaeeea6de3d0e0dd892a5a"} Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.545467 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" event={"ID":"b15da8e1-2f29-41fa-8dae-dce111e40262","Type":"ContainerStarted","Data":"7b55d99eb74589b2033f67cd7863aec6c07c315cdf1ccb3cca565a50c41abf1c"} Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.546425 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.548394 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" event={"ID":"86dbc349-8d0e-4524-a7a6-0292689ad46c","Type":"ContainerStarted","Data":"92caa6f5187aed4dbb31dd4d3c3ac05e3f00fd7235dbce77dbfbd6815074a2cb"} Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.552670 4876 generic.go:334] "Generic (PLEG): container finished" podID="dbb7d698-fe25-4fe1-8207-67bd7af52f7d" containerID="9670438bf0c02843df8961eb8ad9f4fb93b76c2528b37716deb441309c54bb02" exitCode=0 Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.552707 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" event={"ID":"dbb7d698-fe25-4fe1-8207-67bd7af52f7d","Type":"ContainerDied","Data":"9670438bf0c02843df8961eb8ad9f4fb93b76c2528b37716deb441309c54bb02"} Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.567783 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.567990 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.067969274 +0000 UTC m=+138.556633896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.568091 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.568388 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.068376254 +0000 UTC m=+138.557040876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.580750 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.643593 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tx9gn" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.651018 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.659424 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-r7wwl" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.670765 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.671484 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.171456141 +0000 UTC m=+138.660120843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.685495 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.687601 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk"] Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.688920 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c"] Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.739156 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx"] Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.774860 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.775287 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.275274958 +0000 UTC m=+138.763939580 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: W1205 08:17:53.858889 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56f14ba6_fcaf_45fa_84d8_9a4461dda473.slice/crio-edc7bb1510a46c1cf856009f2987d829c77d172b6311e2c685493eac3c9e5a2d WatchSource:0}: Error finding container edc7bb1510a46c1cf856009f2987d829c77d172b6311e2c685493eac3c9e5a2d: Status 404 returned error can't find the container with id edc7bb1510a46c1cf856009f2987d829c77d172b6311e2c685493eac3c9e5a2d Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.877461 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.878064 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.377934265 +0000 UTC m=+138.866598887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:53 crc kubenswrapper[4876]: W1205 08:17:53.892946 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c16d20e_2b84_4436_8c6d_1bf4cea72706.slice/crio-3e21ce2e470ac940d669affef451a9cab106d5f8cebaa38b5aa8a6e84f40af3e WatchSource:0}: Error finding container 3e21ce2e470ac940d669affef451a9cab106d5f8cebaa38b5aa8a6e84f40af3e: Status 404 returned error can't find the container with id 3e21ce2e470ac940d669affef451a9cab106d5f8cebaa38b5aa8a6e84f40af3e Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.928046 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-r8p7s"] Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.936816 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gj226"] Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.939454 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pqx48"] Dec 05 08:17:53 crc kubenswrapper[4876]: I1205 08:17:53.980274 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:53 crc kubenswrapper[4876]: E1205 08:17:53.982664 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.482650633 +0000 UTC m=+138.971315255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.087490 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.087929 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.587915616 +0000 UTC m=+139.076580238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.189075 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.189421 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.689404943 +0000 UTC m=+139.178069565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.290607 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.291066 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.791050525 +0000 UTC m=+139.279715147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.391664 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.391973 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.891961308 +0000 UTC m=+139.380625930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.492918 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.493649 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:54.99362035 +0000 UTC m=+139.482284972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.579654 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" event={"ID":"808f78c5-f306-4f4a-a823-7e114bb9be9f","Type":"ContainerStarted","Data":"f60b0ef24ef10d489da628b905876d4588a20a5007893a83e779ef1922944abe"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.598559 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-6rtbf" event={"ID":"fed5f23f-513b-449f-adaa-d65ce9beb793","Type":"ContainerStarted","Data":"3695dc259be728dbed8f32d52570cecaee7bd0b27de9c7580f994e5eb4c79b34"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.598768 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-6rtbf" event={"ID":"fed5f23f-513b-449f-adaa-d65ce9beb793","Type":"ContainerStarted","Data":"5c544028286709a509b41a096f9103c6a7f57d074ea94c828dac4fe9b61e5d87"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.599283 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.599607 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.09959366 +0000 UTC m=+139.588258282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.614403 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-r7wwl" event={"ID":"56f14ba6-fcaf-45fa-84d8-9a4461dda473","Type":"ContainerStarted","Data":"559a478329edab9c2bce97353e7757f88d8dd0202eb0f497bed6cab4a1926974"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.614451 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-r7wwl" event={"ID":"56f14ba6-fcaf-45fa-84d8-9a4461dda473","Type":"ContainerStarted","Data":"edc7bb1510a46c1cf856009f2987d829c77d172b6311e2c685493eac3c9e5a2d"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.642909 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" event={"ID":"859647d8-54ea-4892-ae54-2179375e1ae0","Type":"ContainerStarted","Data":"0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.644850 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.653565 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" event={"ID":"646287d9-6646-42a6-a5e2-9100f5b9b680","Type":"ContainerStarted","Data":"2b25bb2baaaeb5f172c2c1a3288cc7c532dd810a18dd89d22297390632157f18"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.673570 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" event={"ID":"b15da8e1-2f29-41fa-8dae-dce111e40262","Type":"ContainerStarted","Data":"5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.674381 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.689682 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" event={"ID":"4c5a2919-0f9c-4f28-b6e4-ab288f991483","Type":"ContainerStarted","Data":"462adc33e726ca1803a13411c15090d226d49bb3f43eceb7128a59801a3b54dd"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.694398 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" event={"ID":"36087227-ee02-4dd0-b801-2ced90e4173f","Type":"ContainerStarted","Data":"afe8123cbc3997d749110d179145b97f73c22aa4e993877ef7cc44e202248aeb"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.694685 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.698874 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" event={"ID":"dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf","Type":"ContainerStarted","Data":"945461a53cfe1993c16ffd3c42c98ec364212c64f4c2dad59c3848ec390c0f89"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.700561 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.701889 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.201862647 +0000 UTC m=+139.690527289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.703484 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" event={"ID":"ba874267-c442-44a8-9ed6-c8e0b1bc7505","Type":"ContainerStarted","Data":"f8cdd46a346edee4f790aa87a2156b63bcaefc38a8eb66326b5c0a64ed9a1d8a"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.703508 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" event={"ID":"ba874267-c442-44a8-9ed6-c8e0b1bc7505","Type":"ContainerStarted","Data":"b98532427f1dad0be8a69712293b02c64971d1ca80543619f01e7f8b4b217ebf"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.703607 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.709629 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" event={"ID":"86dbc349-8d0e-4524-a7a6-0292689ad46c","Type":"ContainerStarted","Data":"132d3d18d43e111b4b8836d9a052935fffcb5dcd6f5126cd9c54389fbcdb6b8b"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.712636 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" event={"ID":"2c16d20e-2b84-4436-8c6d-1bf4cea72706","Type":"ContainerStarted","Data":"3e21ce2e470ac940d669affef451a9cab106d5f8cebaa38b5aa8a6e84f40af3e"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.713999 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" event={"ID":"dbb7d698-fe25-4fe1-8207-67bd7af52f7d","Type":"ContainerStarted","Data":"d72bff75eb35e37ee34d46dc2a1fd0838e95f4af78ad86c711c79279c8251ccd"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.714369 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.719933 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" event={"ID":"62c18aef-7ad2-4616-beb0-d286f175f1d1","Type":"ContainerStarted","Data":"6e6e8d469e6e4add4191c69e68b29a51615829da63579250120b1a01a7070393"} Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.720029 4876 patch_prober.go:28] interesting pod/downloads-7954f5f757-vqw7m container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.720070 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vqw7m" podUID="5d3b7e9c-38f3-435a-a765-c9c2f3b360e8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.759619 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" podStartSLOduration=120.759603106 podStartE2EDuration="2m0.759603106s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:54.759102104 +0000 UTC m=+139.247766726" watchObservedRunningTime="2025-12-05 08:17:54.759603106 +0000 UTC m=+139.248267728" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.803321 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.805011 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.304997486 +0000 UTC m=+139.793662108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.811233 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-6rtbf" podStartSLOduration=120.811217562 podStartE2EDuration="2m0.811217562s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:54.810249328 +0000 UTC m=+139.298913950" watchObservedRunningTime="2025-12-05 08:17:54.811217562 +0000 UTC m=+139.299882184" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.862383 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" podStartSLOduration=120.862367566 podStartE2EDuration="2m0.862367566s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:54.86012004 +0000 UTC m=+139.348784662" watchObservedRunningTime="2025-12-05 08:17:54.862367566 +0000 UTC m=+139.351032188" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.911378 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.911618 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.411594542 +0000 UTC m=+139.900259154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.912145 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:54 crc kubenswrapper[4876]: E1205 08:17:54.912504 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.412490434 +0000 UTC m=+139.901155056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.921351 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sch8c" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.927281 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vqw7m" podStartSLOduration=120.927267485 podStartE2EDuration="2m0.927267485s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:54.884265716 +0000 UTC m=+139.372930338" watchObservedRunningTime="2025-12-05 08:17:54.927267485 +0000 UTC m=+139.415932097" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.930276 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.972644 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-r7wwl" podStartSLOduration=4.972620233 podStartE2EDuration="4.972620233s" podCreationTimestamp="2025-12-05 08:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:54.930541437 +0000 UTC m=+139.419206059" watchObservedRunningTime="2025-12-05 08:17:54.972620233 +0000 UTC m=+139.461284855" Dec 05 08:17:54 crc kubenswrapper[4876]: I1205 08:17:54.975327 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" podStartSLOduration=120.9753138 podStartE2EDuration="2m0.9753138s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:54.972269294 +0000 UTC m=+139.460933916" watchObservedRunningTime="2025-12-05 08:17:54.9753138 +0000 UTC m=+139.463978422" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.013001 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.013277 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.513263903 +0000 UTC m=+140.001928525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.065626 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-w85pf" podStartSLOduration=121.065610067 podStartE2EDuration="2m1.065610067s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.064791826 +0000 UTC m=+139.553456458" watchObservedRunningTime="2025-12-05 08:17:55.065610067 +0000 UTC m=+139.554274679" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.066009 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" podStartSLOduration=121.066002907 podStartE2EDuration="2m1.066002907s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.012469103 +0000 UTC m=+139.501133725" watchObservedRunningTime="2025-12-05 08:17:55.066002907 +0000 UTC m=+139.554667529" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.123801 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.124337 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.624317241 +0000 UTC m=+140.112981903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.227179 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.229471 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7nmrr" podStartSLOduration=121.229458 podStartE2EDuration="2m1.229458s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.171602248 +0000 UTC m=+139.660266870" watchObservedRunningTime="2025-12-05 08:17:55.229458 +0000 UTC m=+139.718122622" Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.230170 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.730140127 +0000 UTC m=+140.218804749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.230743 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.251642 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.305017 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-96pbm"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.328678 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.329078 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.82906683 +0000 UTC m=+140.317731452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.365770 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.429875 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.430282 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:55.930267061 +0000 UTC m=+140.418931683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.516859 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.527347 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:17:55 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:17:55 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:17:55 crc kubenswrapper[4876]: healthz check failed Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.527397 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.534481 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.534842 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.034830235 +0000 UTC m=+140.523494847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.635142 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.635577 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.135562904 +0000 UTC m=+140.624227526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.650834 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-plzs7"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.659827 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.677304 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z6kjz"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.736377 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.736678 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.236666972 +0000 UTC m=+140.725331594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.757753 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" event={"ID":"4c5a2919-0f9c-4f28-b6e4-ab288f991483","Type":"ContainerStarted","Data":"afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.757946 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.764470 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" event={"ID":"36087227-ee02-4dd0-b801-2ced90e4173f","Type":"ContainerStarted","Data":"bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.765220 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.768281 4876 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gj226 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.768313 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.769577 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" event={"ID":"a29c5fbc-f737-48f5-87c8-7c7bebc8648d","Type":"ContainerStarted","Data":"bf0049f0430aef77b788ccb0a258fc4e9ffc21e9abe547814ce6f34178243ce0"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.774600 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" event={"ID":"808f78c5-f306-4f4a-a823-7e114bb9be9f","Type":"ContainerStarted","Data":"5f5222c0a008eaa6acb49da95c12873e41e7ecf09c04e9bfbe345f5c735ca074"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.774640 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" event={"ID":"808f78c5-f306-4f4a-a823-7e114bb9be9f","Type":"ContainerStarted","Data":"cda912b30ddd55cc64bf311fe84e3bd87d66bfa081b566f7111b0a4c7c46c0d6"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.783116 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" event={"ID":"2c16d20e-2b84-4436-8c6d-1bf4cea72706","Type":"ContainerStarted","Data":"98a0b1c560bf003586e39e0c3deaa6805b729caeb67c7cadb0c9bdc5cf61fcd9"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.783988 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.785239 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" event={"ID":"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4","Type":"ContainerStarted","Data":"1bc9f8edf05072a37ac99fb13bfc1b248d54764a22661cc2d65aff0b037fa4a5"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.786406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" event={"ID":"6b11e75d-3346-4570-8266-59223cbb0619","Type":"ContainerStarted","Data":"27a888fed469fa02f461b0304f121670b7494fdc6ed009b4fd90f80cbf54c15d"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.786435 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" event={"ID":"6b11e75d-3346-4570-8266-59223cbb0619","Type":"ContainerStarted","Data":"ea152a4b595adc6d4fb4ae40358f0c853a6c9a876d7e467d5c38c79034517f7a"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.792807 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.807196 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" podStartSLOduration=122.807180032 podStartE2EDuration="2m2.807180032s" podCreationTimestamp="2025-12-05 08:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.792702239 +0000 UTC m=+140.281366861" watchObservedRunningTime="2025-12-05 08:17:55.807180032 +0000 UTC m=+140.295844654" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.808630 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.810523 4876 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nhprd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.810571 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" podUID="6b11e75d-3346-4570-8266-59223cbb0619" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.812488 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" event={"ID":"9986f6ed-67ea-4ac8-bae1-0e782bd83112","Type":"ContainerStarted","Data":"9ffd4c19668c8f95f1c644fa7b8b37a1a4e01bf1d46a640d77fe7c3eb2f54e92"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.812779 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" event={"ID":"9986f6ed-67ea-4ac8-bae1-0e782bd83112","Type":"ContainerStarted","Data":"bf669d98418628c7130bc5e7ba7f371008938cf0b81c5d66edb562f0eb1657e9"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.813700 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8czl9"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.815379 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.815687 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.837385 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.838426 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.338410876 +0000 UTC m=+140.827075488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.848010 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dcfx" podStartSLOduration=121.847992877 podStartE2EDuration="2m1.847992877s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.82463103 +0000 UTC m=+140.313295652" watchObservedRunningTime="2025-12-05 08:17:55.847992877 +0000 UTC m=+140.336657499" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.848174 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" event={"ID":"86dbc349-8d0e-4524-a7a6-0292689ad46c","Type":"ContainerStarted","Data":"c50766dd6b9cbb6e31d0b4b35758206fa4f9cb2ae07a6575c07aa3dc728193d3"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.849432 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mm2m6"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.856126 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.857504 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" event={"ID":"81bf7719-efce-4d93-8e77-6273ff226a5e","Type":"ContainerStarted","Data":"c077cd51a16333f19ca74e64a535a4bc249f5821d30923019ea34a4bcf198df7"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.857545 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" event={"ID":"81bf7719-efce-4d93-8e77-6273ff226a5e","Type":"ContainerStarted","Data":"31c6880ada223d87ab46f359667f3d2434141001726e2e63fabf4192a319acd3"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.874761 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.875228 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zc7rw"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.875375 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" podStartSLOduration=121.875366104 podStartE2EDuration="2m1.875366104s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.853223328 +0000 UTC m=+140.341887950" watchObservedRunningTime="2025-12-05 08:17:55.875366104 +0000 UTC m=+140.364030726" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.876567 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" event={"ID":"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160","Type":"ContainerStarted","Data":"a281fa37dac173694a05c7e7418c3098835c87442ecb50679c79e61312519148"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.876600 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" event={"ID":"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160","Type":"ContainerStarted","Data":"c8009aa42756839215d7d11d9fedec6ee2e785c82f949ae05fb1082113ba83b0"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.882727 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" event={"ID":"f0573958-5bcc-4deb-b42b-e2d65ea550b5","Type":"ContainerStarted","Data":"37101902a7882ed76a45c71ed4afe0e93a2809554704280958e301f0d6fe8698"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.900868 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-plzs7" event={"ID":"75f4a218-73a6-4a91-a8ab-6177ee8d1e45","Type":"ContainerStarted","Data":"18860c21b3f9fe8edfdd6d0decf3e6b1c728cfe43626cbb36a6fed4c6f20c660"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.907629 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rd9lr"] Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.914760 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" podStartSLOduration=121.914740192 podStartE2EDuration="2m1.914740192s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.885302713 +0000 UTC m=+140.373967335" watchObservedRunningTime="2025-12-05 08:17:55.914740192 +0000 UTC m=+140.403404814" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.943860 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gfdrk" podStartSLOduration=121.943840923 podStartE2EDuration="2m1.943840923s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.902659389 +0000 UTC m=+140.391324011" watchObservedRunningTime="2025-12-05 08:17:55.943840923 +0000 UTC m=+140.432505545" Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.945241 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" event={"ID":"646287d9-6646-42a6-a5e2-9100f5b9b680","Type":"ContainerStarted","Data":"a782323c7b3771cbc39369c16f5da6af6c055369a6587ca6b47143b852e1ccf8"} Dec 05 08:17:55 crc kubenswrapper[4876]: I1205 08:17:55.963474 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:55 crc kubenswrapper[4876]: E1205 08:17:55.973666 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.473638131 +0000 UTC m=+140.962302753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.003003 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vzhwk"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.004345 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.007436 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.014989 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-4cv77" podStartSLOduration=122.014966718 podStartE2EDuration="2m2.014966718s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.926729103 +0000 UTC m=+140.415393725" watchObservedRunningTime="2025-12-05 08:17:56.014966718 +0000 UTC m=+140.503631340" Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.028789 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-96pbm" podStartSLOduration=122.028772245 podStartE2EDuration="2m2.028772245s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:55.985769035 +0000 UTC m=+140.474433647" watchObservedRunningTime="2025-12-05 08:17:56.028772245 +0000 UTC m=+140.517436867" Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.029268 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.039953 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.058711 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c25rq" podStartSLOduration=123.058688996 podStartE2EDuration="2m3.058688996s" podCreationTimestamp="2025-12-05 08:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:56.03974006 +0000 UTC m=+140.528404682" watchObservedRunningTime="2025-12-05 08:17:56.058688996 +0000 UTC m=+140.547353618" Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.058924 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tx9gn"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.065504 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.067967 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.566861191 +0000 UTC m=+141.055525813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.073963 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.077756 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.577742454 +0000 UTC m=+141.066407076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: W1205 08:17:56.081579 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06abb292_ed23_4de1_9974_c487e7bb2933.slice/crio-681ecb9707270e7ff6cb0f4d83fee010cebec3fe7301c79d36fbb7b6b3184d33 WatchSource:0}: Error finding container 681ecb9707270e7ff6cb0f4d83fee010cebec3fe7301c79d36fbb7b6b3184d33: Status 404 returned error can't find the container with id 681ecb9707270e7ff6cb0f4d83fee010cebec3fe7301c79d36fbb7b6b3184d33 Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.085228 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.098385 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xb5cw"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.103512 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-scljl"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.111247 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc"] Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.143430 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.178455 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.178792 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.67877364 +0000 UTC m=+141.167438262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: W1205 08:17:56.209011 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cdca41c_e056_44f6_9f06_6c129d1368cd.slice/crio-ccc4a355d7611352a5968d92c1fb6b716fef00bf4a1ba1fe5688599290891969 WatchSource:0}: Error finding container ccc4a355d7611352a5968d92c1fb6b716fef00bf4a1ba1fe5688599290891969: Status 404 returned error can't find the container with id ccc4a355d7611352a5968d92c1fb6b716fef00bf4a1ba1fe5688599290891969 Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.249501 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-pqx48" podStartSLOduration=123.249484795 podStartE2EDuration="2m3.249484795s" podCreationTimestamp="2025-12-05 08:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:56.222055226 +0000 UTC m=+140.710719848" watchObservedRunningTime="2025-12-05 08:17:56.249484795 +0000 UTC m=+140.738149417" Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.281116 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.281380 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.781370005 +0000 UTC m=+141.270034627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.384499 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.388762 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.888745761 +0000 UTC m=+141.377410383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.493616 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.494189 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:56.994177607 +0000 UTC m=+141.482842229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.522352 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:17:56 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:17:56 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:17:56 crc kubenswrapper[4876]: healthz check failed Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.522402 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.597243 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.597473 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.09745931 +0000 UTC m=+141.586123932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.597754 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.598115 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.098108236 +0000 UTC m=+141.586772858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.698881 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.699483 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.199468101 +0000 UTC m=+141.688132733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.800967 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.801294 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.301283016 +0000 UTC m=+141.789947638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:56 crc kubenswrapper[4876]: I1205 08:17:56.905012 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:56 crc kubenswrapper[4876]: E1205 08:17:56.905686 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.405670537 +0000 UTC m=+141.894335159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.006831 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.007149 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.507137944 +0000 UTC m=+141.995802566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.032051 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" event={"ID":"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd","Type":"ContainerStarted","Data":"dc72dcb557813776518f855e39a4b626d35ed35ba800faa78c8edb6fd2657f4b"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.077663 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-plzs7" event={"ID":"75f4a218-73a6-4a91-a8ab-6177ee8d1e45","Type":"ContainerStarted","Data":"223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.090659 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" event={"ID":"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4","Type":"ContainerStarted","Data":"c011656bdc4ed76a7f8cf51b0a26e44cb8a036337a6d51d991be6709b664987c"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.090701 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" event={"ID":"43d50654-e5bb-495a-8c3e-89bbb7dc3bf4","Type":"ContainerStarted","Data":"ddbe4ca3d1dc00e849347e788a3bc2898acb014727fab605b3db8df38a3154f1"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.103434 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" event={"ID":"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c","Type":"ContainerStarted","Data":"1196966717f162042fe6d2c800deb55dc12014c32d537c44f58a4b8c0a1d3575"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.103469 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" event={"ID":"5d4482e5-4766-4e46-a4a4-f02af8e4ac1c","Type":"ContainerStarted","Data":"c820559ab3fe00621c6d1f6d4961206d9f243a87d1cb09cd5c70dc64b87c8e5a"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.107919 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.108467 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.608454387 +0000 UTC m=+142.097119009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.114630 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-plzs7" podStartSLOduration=123.114620042 podStartE2EDuration="2m3.114620042s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.111481823 +0000 UTC m=+141.600146445" watchObservedRunningTime="2025-12-05 08:17:57.114620042 +0000 UTC m=+141.603284664" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.135529 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" event={"ID":"92e719e9-c64e-4e54-a002-427ebe228761","Type":"ContainerStarted","Data":"8aa08102288855b46df9243b66f38cd0e94a901ac86c2301ac84d0ec08b60c3f"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.150368 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" event={"ID":"f0573958-5bcc-4deb-b42b-e2d65ea550b5","Type":"ContainerStarted","Data":"e1527758881d580fbbf16fc57609ae76e18be4041cf35866262b8f53e31f3f93"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.156221 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rd9lr" event={"ID":"eece9379-11d6-44ff-9ca9-11ef43acad82","Type":"ContainerStarted","Data":"993471579f197151f4b688e24373934cc3c0dcd98f4a77c63d3660d1873fda34"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.157654 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" event={"ID":"5cdca41c-e056-44f6-9f06-6c129d1368cd","Type":"ContainerStarted","Data":"ccc4a355d7611352a5968d92c1fb6b716fef00bf4a1ba1fe5688599290891969"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.159048 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" event={"ID":"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8","Type":"ContainerStarted","Data":"20e7dcab8015f200c5145730b2b4308454a65e72166726259f6d0531c62f0a5c"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.164762 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pb7jc" podStartSLOduration=123.16474904 podStartE2EDuration="2m3.16474904s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.163312934 +0000 UTC m=+141.651977556" watchObservedRunningTime="2025-12-05 08:17:57.16474904 +0000 UTC m=+141.653413662" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.170811 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-cmsv8" podStartSLOduration=123.170793372 podStartE2EDuration="2m3.170793372s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.136621334 +0000 UTC m=+141.625285956" watchObservedRunningTime="2025-12-05 08:17:57.170793372 +0000 UTC m=+141.659457994" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.174275 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" event={"ID":"c6a5866e-5650-4e91-9003-3d7537713d05","Type":"ContainerStarted","Data":"567c7809ba40b4e7e3481b9e5fc7815b3ca2378ca13ea6d8a90e59a7732d2831"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.196586 4876 generic.go:334] "Generic (PLEG): container finished" podID="a29c5fbc-f737-48f5-87c8-7c7bebc8648d" containerID="4f9b7a73cac26dc76ef8297041d954db9a5dfaf8fd12ad18eb0e3c085e276345" exitCode=0 Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.197279 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" event={"ID":"a29c5fbc-f737-48f5-87c8-7c7bebc8648d","Type":"ContainerDied","Data":"4f9b7a73cac26dc76ef8297041d954db9a5dfaf8fd12ad18eb0e3c085e276345"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.204016 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-z6kjz" podStartSLOduration=123.204000706 podStartE2EDuration="2m3.204000706s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.203415161 +0000 UTC m=+141.692079793" watchObservedRunningTime="2025-12-05 08:17:57.204000706 +0000 UTC m=+141.692665328" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.205910 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" event={"ID":"ed2fd37f-5667-4509-a422-92dba00024c7","Type":"ContainerStarted","Data":"48d6281dc34a9e351c52c9bb9884623e93648da87479cec30e2c748e9f0e9797"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.209627 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.211440 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.711415202 +0000 UTC m=+142.200079824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.213656 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" event={"ID":"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f","Type":"ContainerStarted","Data":"c1c8ad67e120c07317267ac985079983cd40a60158032a7a7930f510c52afdfc"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.224456 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kdhx2"] Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.225335 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.228182 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.248190 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" event={"ID":"4d8130b4-8e6a-462c-bc65-d58567c00cd1","Type":"ContainerStarted","Data":"d8def0a096627c35f8e863e875496225f60eca6be601436befaf28ef17c9342d"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.248231 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" event={"ID":"4d8130b4-8e6a-462c-bc65-d58567c00cd1","Type":"ContainerStarted","Data":"27a069e9cfbd78992a3e37db4ca26bf3ce7a8bff1a9a09872e6d5ec9d9d0e463"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.253152 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kdhx2"] Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.259293 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tx9gn" event={"ID":"1d2073f7-877e-4ba2-bc9b-a8d34d15a447","Type":"ContainerStarted","Data":"309aebf78877bf1c597b37b444f1d9ef9222258b3d0d7374a406972da7e74ddf"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.259345 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tx9gn" event={"ID":"1d2073f7-877e-4ba2-bc9b-a8d34d15a447","Type":"ContainerStarted","Data":"dea0eb9e1a02ed122772883fcd60410a64f9840c74d9faafb8731805fcfe492a"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.266393 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8czl9" event={"ID":"6c9db69d-c1c4-4326-abb1-16d92bd6441d","Type":"ContainerStarted","Data":"c8fd09de0cb7ac67565ad8c714c979f712d8ae08b131aaf85825be91a7305cb9"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.266460 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8czl9" event={"ID":"6c9db69d-c1c4-4326-abb1-16d92bd6441d","Type":"ContainerStarted","Data":"b17b8e79660418fe4ac80c39cc0cead145e00a8b9ea1da8b5b01af68c0df738d"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.267002 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.271437 4876 patch_prober.go:28] interesting pod/console-operator-58897d9998-8czl9 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/readyz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.271484 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8czl9" podUID="6c9db69d-c1c4-4326-abb1-16d92bd6441d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/readyz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.277116 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" event={"ID":"c4ffecba-7d0e-42ca-8b18-9792a7ef34a4","Type":"ContainerStarted","Data":"0deb76bc1b4dd4472da5ce32e86109f479d22e7706021cb5afca9f6d2f333014"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.278056 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-tx9gn" podStartSLOduration=7.278039184 podStartE2EDuration="7.278039184s" podCreationTimestamp="2025-12-05 08:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.277255794 +0000 UTC m=+141.765920416" watchObservedRunningTime="2025-12-05 08:17:57.278039184 +0000 UTC m=+141.766703806" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.298451 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-8czl9" podStartSLOduration=123.298438366 podStartE2EDuration="2m3.298438366s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.298067607 +0000 UTC m=+141.786732229" watchObservedRunningTime="2025-12-05 08:17:57.298438366 +0000 UTC m=+141.787102988" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.313648 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.313819 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdlbc\" (UniqueName: \"kubernetes.io/projected/dfa346f8-9447-4712-b791-21bf718356b3-kube-api-access-kdlbc\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.313957 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-utilities\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.313977 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-catalog-content\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.314674 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.814659743 +0000 UTC m=+142.303324365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.323621 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" event={"ID":"e2cf0d42-b34d-4f7a-b1b9-d1b92eb9e160","Type":"ContainerStarted","Data":"711be5ed9e43fd5298ae320ba2f82355c6173ae5405d1d34ba5c5c76f44e4e7e"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.332125 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" event={"ID":"9e452b45-5f78-4f52-b96e-83aa3dee3376","Type":"ContainerStarted","Data":"f4fbaeedb242144d707913789e9b0cef3980698f537865a190d33f774acddf40"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.334375 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lrlj" podStartSLOduration=123.334343667 podStartE2EDuration="2m3.334343667s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.329161077 +0000 UTC m=+141.817825699" watchObservedRunningTime="2025-12-05 08:17:57.334343667 +0000 UTC m=+141.823008299" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.341345 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" event={"ID":"b5742635-f4b3-47ee-88be-d59720792b25","Type":"ContainerStarted","Data":"efb2cedad915817582bbb028575ecd4ad526db2eb7c49cf2eb674d1c19766ef6"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.341389 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" event={"ID":"b5742635-f4b3-47ee-88be-d59720792b25","Type":"ContainerStarted","Data":"887c6fb4a15e34bd8de607adcbfe43b0635b059582027c82b1ea1ba099163e74"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.359411 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bhgfr" podStartSLOduration=123.359395976 podStartE2EDuration="2m3.359395976s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.35875243 +0000 UTC m=+141.847417052" watchObservedRunningTime="2025-12-05 08:17:57.359395976 +0000 UTC m=+141.848060598" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.363408 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" event={"ID":"8eaba618-88b4-4d8d-9b87-fdabef6b7aad","Type":"ContainerStarted","Data":"20e54b836c661fa4476bd9eb45debdeca108136612dea6b3439b9868b62905c3"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.363449 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" event={"ID":"8eaba618-88b4-4d8d-9b87-fdabef6b7aad","Type":"ContainerStarted","Data":"937a81c921ce9e5140eb21fef0fcd8bcb4a07a132c7f19abadc48e4edca42129"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.369531 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" event={"ID":"06abb292-ed23-4de1-9974-c487e7bb2933","Type":"ContainerStarted","Data":"681ecb9707270e7ff6cb0f4d83fee010cebec3fe7301c79d36fbb7b6b3184d33"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.380027 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" event={"ID":"07af6910-37e6-401a-a863-c3de22d0e39f","Type":"ContainerStarted","Data":"4cb217edcc351ba45f34f91d72730fb28e9f7f42199aa89544bb7ab6b0a6d382"} Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.380651 4876 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gj226 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.380706 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.400010 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhprd" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.409486 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" podStartSLOduration=123.409467173 podStartE2EDuration="2m3.409467173s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.391342438 +0000 UTC m=+141.880007060" watchObservedRunningTime="2025-12-05 08:17:57.409467173 +0000 UTC m=+141.898131795" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.415852 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdlbc\" (UniqueName: \"kubernetes.io/projected/dfa346f8-9447-4712-b791-21bf718356b3-kube-api-access-kdlbc\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.416618 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.416786 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-utilities\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.417002 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-catalog-content\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.417822 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-utilities\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.418027 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:57.918013778 +0000 UTC m=+142.406678400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.418389 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-catalog-content\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.432894 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbspv" podStartSLOduration=123.432876081 podStartE2EDuration="2m3.432876081s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.411519155 +0000 UTC m=+141.900183777" watchObservedRunningTime="2025-12-05 08:17:57.432876081 +0000 UTC m=+141.921540703" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.434081 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nnc6w"] Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.434886 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.443139 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.475658 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nnc6w"] Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.476129 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" podStartSLOduration=123.476111006 podStartE2EDuration="2m3.476111006s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:57.456726929 +0000 UTC m=+141.945391551" watchObservedRunningTime="2025-12-05 08:17:57.476111006 +0000 UTC m=+141.964775628" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.477576 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdlbc\" (UniqueName: \"kubernetes.io/projected/dfa346f8-9447-4712-b791-21bf718356b3-kube-api-access-kdlbc\") pod \"community-operators-kdhx2\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.511636 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:17:57 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:17:57 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:17:57 crc kubenswrapper[4876]: healthz check failed Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.511946 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.518330 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.518673 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-utilities\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.518842 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrspc\" (UniqueName: \"kubernetes.io/projected/17ce8bf7-9544-4849-aa82-c5744a929656-kube-api-access-mrspc\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.518885 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-catalog-content\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.519917 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.019888285 +0000 UTC m=+142.508552907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.563082 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.620377 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrspc\" (UniqueName: \"kubernetes.io/projected/17ce8bf7-9544-4849-aa82-c5744a929656-kube-api-access-mrspc\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.620422 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-catalog-content\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.620480 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.620514 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-utilities\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.620838 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-utilities\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.621315 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-catalog-content\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.621556 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.121545837 +0000 UTC m=+142.610210459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.649293 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vtmg6"] Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.650890 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.668243 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vtmg6"] Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.699927 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrspc\" (UniqueName: \"kubernetes.io/projected/17ce8bf7-9544-4849-aa82-c5744a929656-kube-api-access-mrspc\") pod \"certified-operators-nnc6w\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.721568 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.722125 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.222108641 +0000 UTC m=+142.710773263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.814241 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-24njp"] Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.826773 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b7r8\" (UniqueName: \"kubernetes.io/projected/69511d2e-da37-424d-bb34-4791b3292c35-kube-api-access-4b7r8\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.826813 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-utilities\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.826840 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-catalog-content\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.826918 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.827231 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.32721972 +0000 UTC m=+142.815884342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.838397 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.850340 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24njp"] Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.859342 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.936617 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.936891 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b7r8\" (UniqueName: \"kubernetes.io/projected/69511d2e-da37-424d-bb34-4791b3292c35-kube-api-access-4b7r8\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.936942 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-utilities\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.936958 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-catalog-content\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.936991 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-catalog-content\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.937035 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-utilities\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.937072 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjqqb\" (UniqueName: \"kubernetes.io/projected/d0e342ab-27c1-442c-926e-f2e9ed1153c5-kube-api-access-zjqqb\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:57 crc kubenswrapper[4876]: E1205 08:17:57.937193 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.43717706 +0000 UTC m=+142.925841682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.937809 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-utilities\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.938979 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-catalog-content\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:57 crc kubenswrapper[4876]: I1205 08:17:57.982162 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b7r8\" (UniqueName: \"kubernetes.io/projected/69511d2e-da37-424d-bb34-4791b3292c35-kube-api-access-4b7r8\") pod \"community-operators-vtmg6\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.047614 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjqqb\" (UniqueName: \"kubernetes.io/projected/d0e342ab-27c1-442c-926e-f2e9ed1153c5-kube-api-access-zjqqb\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.047928 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-catalog-content\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.047962 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.047985 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-utilities\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.048372 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-utilities\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.048818 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-catalog-content\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.049067 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.549055358 +0000 UTC m=+143.037719980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.089724 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjqqb\" (UniqueName: \"kubernetes.io/projected/d0e342ab-27c1-442c-926e-f2e9ed1153c5-kube-api-access-zjqqb\") pod \"certified-operators-24njp\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.100451 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.145331 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h9vc6" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.151400 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.151918 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.651865469 +0000 UTC m=+143.140530091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.251157 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.252751 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.253085 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.75307404 +0000 UTC m=+143.241738662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.285616 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kdhx2"] Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.354364 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.354725 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.854710571 +0000 UTC m=+143.343375193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.401729 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" event={"ID":"9e452b45-5f78-4f52-b96e-83aa3dee3376","Type":"ContainerStarted","Data":"d053cba2d1d227d48c3c2f2eea25190bc3bdf6ffa99f524205e871e75fe2361c"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.401974 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" event={"ID":"9e452b45-5f78-4f52-b96e-83aa3dee3376","Type":"ContainerStarted","Data":"aa018afb8880c6cf4462d65d750861671df5233de0742b9aeb5c8348980a3c71"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.402749 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.439167 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" podStartSLOduration=124.43914344 podStartE2EDuration="2m4.43914344s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.432528164 +0000 UTC m=+142.921192786" watchObservedRunningTime="2025-12-05 08:17:58.43914344 +0000 UTC m=+142.927808062" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.443730 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nnc6w"] Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.444130 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" event={"ID":"4d8130b4-8e6a-462c-bc65-d58567c00cd1","Type":"ContainerStarted","Data":"e94022cb79ab1aa2092683bdd2563b6c7df5ebe8d3a7abfb06750fe703a8fd1a"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.456889 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.457226 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:58.957214304 +0000 UTC m=+143.445878916 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.462111 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" event={"ID":"06abb292-ed23-4de1-9974-c487e7bb2933","Type":"ContainerStarted","Data":"d546cc1c2517b59f0398f7721859586f061c9338284a381bafa5fad147199c53"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.482417 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5jxj6" podStartSLOduration=124.482387306 podStartE2EDuration="2m4.482387306s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.473923193 +0000 UTC m=+142.962587815" watchObservedRunningTime="2025-12-05 08:17:58.482387306 +0000 UTC m=+142.971051928" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.499205 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" event={"ID":"5cdca41c-e056-44f6-9f06-6c129d1368cd","Type":"ContainerStarted","Data":"68342fc2218c422602ddf06d03905127dffc6966d85a244c5d13532e922ba03f"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.499244 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" event={"ID":"5cdca41c-e056-44f6-9f06-6c129d1368cd","Type":"ContainerStarted","Data":"60eda707f8a65c2731c0a3b74dfa8822ee5693a7b494b1d176d58caab0eaac1a"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.508416 4876 generic.go:334] "Generic (PLEG): container finished" podID="92e719e9-c64e-4e54-a002-427ebe228761" containerID="beda37deb799b2caf19bb21252e9bdf3d3ae1fdbcb4ba72fe7121ce6bb12aba3" exitCode=0 Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.508469 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" event={"ID":"92e719e9-c64e-4e54-a002-427ebe228761","Type":"ContainerDied","Data":"beda37deb799b2caf19bb21252e9bdf3d3ae1fdbcb4ba72fe7121ce6bb12aba3"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.515170 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:17:58 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:17:58 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:17:58 crc kubenswrapper[4876]: healthz check failed Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.515205 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.535501 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" event={"ID":"438fd9fa-e2c6-46f4-a54b-1f61cfc9c54f","Type":"ContainerStarted","Data":"0f58650dd0a4da5260bd041ead23df56417c3ff572e3a9623078ec6ea48db6b9"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.537689 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-scljl" podStartSLOduration=124.537671653 podStartE2EDuration="2m4.537671653s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.535759815 +0000 UTC m=+143.024424427" watchObservedRunningTime="2025-12-05 08:17:58.537671653 +0000 UTC m=+143.026336275" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.538766 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfjn" podStartSLOduration=124.53876133 podStartE2EDuration="2m4.53876133s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.506618104 +0000 UTC m=+142.995282726" watchObservedRunningTime="2025-12-05 08:17:58.53876133 +0000 UTC m=+143.027425952" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.559734 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.560026 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.060001343 +0000 UTC m=+143.548665965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.560201 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.561616 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.061606363 +0000 UTC m=+143.550271065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.565061 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdhx2" event={"ID":"dfa346f8-9447-4712-b791-21bf718356b3","Type":"ContainerStarted","Data":"5156aca98613496935a888e7333b99789c09ccfefb4c065b92110f0fc3d230c3"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.604140 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-qbjhf" podStartSLOduration=124.604124371 podStartE2EDuration="2m4.604124371s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.60368803 +0000 UTC m=+143.092352652" watchObservedRunningTime="2025-12-05 08:17:58.604124371 +0000 UTC m=+143.092788993" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.615915 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vtmg6"] Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.622171 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rd9lr" event={"ID":"eece9379-11d6-44ff-9ca9-11ef43acad82","Type":"ContainerStarted","Data":"bb1890ba1fa4b4e29186150023f9bd8bb02af607afe696f5a8f25e52d3827cb5"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.623226 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rd9lr" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.649151 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rd9lr" podStartSLOduration=8.649134651 podStartE2EDuration="8.649134651s" podCreationTimestamp="2025-12-05 08:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.64750347 +0000 UTC m=+143.136168102" watchObservedRunningTime="2025-12-05 08:17:58.649134651 +0000 UTC m=+143.137799273" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.656241 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" event={"ID":"ed2fd37f-5667-4509-a422-92dba00024c7","Type":"ContainerStarted","Data":"b3e52a469b35d83d4f5babfe4b3da5caacc8c0d8965005d271d42a1aa58ee206"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.656289 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" event={"ID":"ed2fd37f-5667-4509-a422-92dba00024c7","Type":"ContainerStarted","Data":"9acabc74513d1b22d71902687cc513e87bcb3b8381de2aa1cd3391e993c09edb"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.663230 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.664406 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.164388154 +0000 UTC m=+143.653052776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.709868 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" event={"ID":"a29c5fbc-f737-48f5-87c8-7c7bebc8648d","Type":"ContainerStarted","Data":"b33ab99a3d7b820fbb12da9d71c21a35e97a638e07376524fc14836f43719b58"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.715248 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-l827b" event={"ID":"07af6910-37e6-401a-a863-c3de22d0e39f","Type":"ContainerStarted","Data":"a830f0ff9201ae143f90ba95a5c61467087e5a46e7a476d2b791f05d342ebd09"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.745538 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g29vs" podStartSLOduration=124.74551509 podStartE2EDuration="2m4.74551509s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.696179492 +0000 UTC m=+143.184844114" watchObservedRunningTime="2025-12-05 08:17:58.74551509 +0000 UTC m=+143.234179712" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.746286 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24njp"] Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.749721 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" event={"ID":"c6a5866e-5650-4e91-9003-3d7537713d05","Type":"ContainerStarted","Data":"7037382b4f8c6db448798cef782f0819b2ac40752be2fa2c4b3dfb2093096e5c"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.749751 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" event={"ID":"c6a5866e-5650-4e91-9003-3d7537713d05","Type":"ContainerStarted","Data":"627868d1fdc95c093fc9a5faf1e1e383cae906e6d42b03c11019bbe4816ae65d"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.754700 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" podStartSLOduration=124.75468147 podStartE2EDuration="2m4.75468147s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.754403083 +0000 UTC m=+143.243067705" watchObservedRunningTime="2025-12-05 08:17:58.75468147 +0000 UTC m=+143.243346092" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.763416 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" event={"ID":"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd","Type":"ContainerStarted","Data":"44f2af726ca15a0558663c445a659946e44438209434e2f116dbec1213be6abb"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.765279 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.787469 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.287431742 +0000 UTC m=+143.776096364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.790427 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" event={"ID":"d726e1ea-2af3-4ade-80cb-e92fe8bd16d8","Type":"ContainerStarted","Data":"af20d1b5dde0cc1e5076e7ad0429adb5acae29cdff9fba0778338cebca4f6e72"} Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.792854 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-mm2m6" podStartSLOduration=124.792838578 podStartE2EDuration="2m4.792838578s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.790111699 +0000 UTC m=+143.278776321" watchObservedRunningTime="2025-12-05 08:17:58.792838578 +0000 UTC m=+143.281503190" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.827100 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.838593 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vzhwk" podStartSLOduration=124.838574266 podStartE2EDuration="2m4.838574266s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:58.817430985 +0000 UTC m=+143.306095607" watchObservedRunningTime="2025-12-05 08:17:58.838574266 +0000 UTC m=+143.327238888" Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.866297 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.867387 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.367366899 +0000 UTC m=+143.856031521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:58 crc kubenswrapper[4876]: I1205 08:17:58.969984 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:58 crc kubenswrapper[4876]: E1205 08:17:58.972294 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.472278332 +0000 UTC m=+143.960943054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.075472 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.076173 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.57615831 +0000 UTC m=+144.064822932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.150976 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-8czl9" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.176859 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.177244 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.677232737 +0000 UTC m=+144.165897359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.278095 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.278298 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.778269653 +0000 UTC m=+144.266934285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.278384 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.278760 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.778749675 +0000 UTC m=+144.267414297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.379518 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.379858 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.879843963 +0000 UTC m=+144.368508585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.411186 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4g9z7"] Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.412285 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.417337 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.427911 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4g9z7"] Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.480973 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.481354 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:17:59.981339201 +0000 UTC m=+144.470003823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.507639 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:17:59 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:17:59 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:17:59 crc kubenswrapper[4876]: healthz check failed Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.507710 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.581536 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.581700 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt7qv\" (UniqueName: \"kubernetes.io/projected/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-kube-api-access-zt7qv\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.581736 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:18:00.08170618 +0000 UTC m=+144.570370852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.581840 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-utilities\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.581879 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.582043 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-catalog-content\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.582234 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:18:00.082221633 +0000 UTC m=+144.570886255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.683047 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.683288 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt7qv\" (UniqueName: \"kubernetes.io/projected/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-kube-api-access-zt7qv\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.683318 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:18:00.18329484 +0000 UTC m=+144.671959462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.683402 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-utilities\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.683440 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.683608 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-catalog-content\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.683801 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:18:00.183791653 +0000 UTC m=+144.672456275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.684027 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-catalog-content\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.684022 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-utilities\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.717682 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt7qv\" (UniqueName: \"kubernetes.io/projected/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-kube-api-access-zt7qv\") pod \"redhat-marketplace-4g9z7\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.727300 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.785187 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.785376 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:18:00.285349892 +0000 UTC m=+144.774014514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.785766 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.786093 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 08:18:00.286086121 +0000 UTC m=+144.774750743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-sklcd" (UID: "907ad872-3805-403d-9649-11e7090bb789") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.786201 4876 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.808053 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" event={"ID":"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd","Type":"ContainerStarted","Data":"4991d128c9e4d41db6c13b7c77801b22c523e9ead9dd83dbae2babcaef2d9827"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.809671 4876 generic.go:334] "Generic (PLEG): container finished" podID="17ce8bf7-9544-4849-aa82-c5744a929656" containerID="a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327" exitCode=0 Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.809758 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnc6w" event={"ID":"17ce8bf7-9544-4849-aa82-c5744a929656","Type":"ContainerDied","Data":"a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.809820 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnc6w" event={"ID":"17ce8bf7-9544-4849-aa82-c5744a929656","Type":"ContainerStarted","Data":"ea1260633b2bca9f2eb6a220eb89ad59090e33b1cd3952ae29841fdd8f5a5cc2"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.811307 4876 generic.go:334] "Generic (PLEG): container finished" podID="dfa346f8-9447-4712-b791-21bf718356b3" containerID="f8a94d614de8aa369b4405d9106cb2ec4e1c9183f5a20d42c74ee0b6998596e5" exitCode=0 Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.811384 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdhx2" event={"ID":"dfa346f8-9447-4712-b791-21bf718356b3","Type":"ContainerDied","Data":"f8a94d614de8aa369b4405d9106cb2ec4e1c9183f5a20d42c74ee0b6998596e5"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.817921 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.818738 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" event={"ID":"92e719e9-c64e-4e54-a002-427ebe228761","Type":"ContainerStarted","Data":"216cbfe258e63dfa3611df09e3a5095e21005f21f5276183bb3cf2338cefd8e2"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.818779 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" event={"ID":"92e719e9-c64e-4e54-a002-427ebe228761","Type":"ContainerStarted","Data":"8b52475f84e14de11bf24dabbeffe485d20989a6ea120ca5d9576a4f194123a5"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.821812 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ms85q"] Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.823059 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.848760 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rd9lr" event={"ID":"eece9379-11d6-44ff-9ca9-11ef43acad82","Type":"ContainerStarted","Data":"4643fafb7dedea6004af0144444480ef0ff1966aa58b9db9ba80d5e01bd5637b"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.857356 4876 generic.go:334] "Generic (PLEG): container finished" podID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerID="4bfe52d0ca7cb917d415116724ed0c1ea0d94c931056e842d45b02e67eca05e6" exitCode=0 Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.857450 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24njp" event={"ID":"d0e342ab-27c1-442c-926e-f2e9ed1153c5","Type":"ContainerDied","Data":"4bfe52d0ca7cb917d415116724ed0c1ea0d94c931056e842d45b02e67eca05e6"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.857472 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24njp" event={"ID":"d0e342ab-27c1-442c-926e-f2e9ed1153c5","Type":"ContainerStarted","Data":"5df63ae6d04ec33016709bede1bbc2fdf09b6bcc57cc913938a1da88ce62f130"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.859063 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ms85q"] Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.887350 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:17:59 crc kubenswrapper[4876]: E1205 08:17:59.887882 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 08:18:00.387857705 +0000 UTC m=+144.876522327 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.891026 4876 generic.go:334] "Generic (PLEG): container finished" podID="69511d2e-da37-424d-bb34-4791b3292c35" containerID="f93c474830bfcb8e4e41b0f1c09f83632caf1d5d2b13e906e487f85f7f3cde07" exitCode=0 Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.892031 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtmg6" event={"ID":"69511d2e-da37-424d-bb34-4791b3292c35","Type":"ContainerDied","Data":"f93c474830bfcb8e4e41b0f1c09f83632caf1d5d2b13e906e487f85f7f3cde07"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.892053 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtmg6" event={"ID":"69511d2e-da37-424d-bb34-4791b3292c35","Type":"ContainerStarted","Data":"58a9a94da32b6650d9c03b97e6073b8547847ae02d9c03a01881880ba58bdcf3"} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.930398 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" podStartSLOduration=125.930373413 podStartE2EDuration="2m5.930373413s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:17:59.904321889 +0000 UTC m=+144.392986511" watchObservedRunningTime="2025-12-05 08:17:59.930373413 +0000 UTC m=+144.419038035" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.984287 4876 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-05T08:17:59.786220744Z","Handler":null,"Name":""} Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.989709 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.989872 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxm4w\" (UniqueName: \"kubernetes.io/projected/e189296a-a3f0-4937-aa28-c9be917fd59c-kube-api-access-cxm4w\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.990062 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-utilities\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.990093 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-catalog-content\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.990450 4876 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.990491 4876 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.996276 4876 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 08:17:59 crc kubenswrapper[4876]: I1205 08:17:59.996313 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.044778 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-sklcd\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.091386 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.091714 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxm4w\" (UniqueName: \"kubernetes.io/projected/e189296a-a3f0-4937-aa28-c9be917fd59c-kube-api-access-cxm4w\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.091755 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-utilities\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.091774 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-catalog-content\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.092333 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-catalog-content\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.092979 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-utilities\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.095714 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.115092 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxm4w\" (UniqueName: \"kubernetes.io/projected/e189296a-a3f0-4937-aa28-c9be917fd59c-kube-api-access-cxm4w\") pod \"redhat-marketplace-ms85q\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.120785 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.168521 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.273995 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4g9z7"] Dec 05 08:18:00 crc kubenswrapper[4876]: W1205 08:18:00.326713 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c1ee83e_b09e_4e3c_b533_6e1ad049eb8e.slice/crio-6d32599322287784390563ecbfa4cc0bc60ce1e10a274c52e19533e2d5bf05b8 WatchSource:0}: Error finding container 6d32599322287784390563ecbfa4cc0bc60ce1e10a274c52e19533e2d5bf05b8: Status 404 returned error can't find the container with id 6d32599322287784390563ecbfa4cc0bc60ce1e10a274c52e19533e2d5bf05b8 Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.431395 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rf92g"] Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.432713 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.437236 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.463116 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rf92g"] Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.506730 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:00 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:00 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:00 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.507101 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.559841 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sklcd"] Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.607955 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-utilities\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.608029 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnqdv\" (UniqueName: \"kubernetes.io/projected/b48955db-dd1c-4e11-8d01-d465e6e7a75d-kube-api-access-cnqdv\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.608054 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-catalog-content\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.665176 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ms85q"] Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.708865 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-utilities\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.708995 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnqdv\" (UniqueName: \"kubernetes.io/projected/b48955db-dd1c-4e11-8d01-d465e6e7a75d-kube-api-access-cnqdv\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.709037 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-catalog-content\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.710161 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-utilities\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.711869 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-catalog-content\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.740789 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnqdv\" (UniqueName: \"kubernetes.io/projected/b48955db-dd1c-4e11-8d01-d465e6e7a75d-kube-api-access-cnqdv\") pod \"redhat-operators-rf92g\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.772328 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.849016 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gfzbx"] Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.851118 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.875561 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gfzbx"] Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.951047 4876 generic.go:334] "Generic (PLEG): container finished" podID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerID="877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f" exitCode=0 Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.951134 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4g9z7" event={"ID":"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e","Type":"ContainerDied","Data":"877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f"} Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.951164 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4g9z7" event={"ID":"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e","Type":"ContainerStarted","Data":"6d32599322287784390563ecbfa4cc0bc60ce1e10a274c52e19533e2d5bf05b8"} Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.958303 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" event={"ID":"907ad872-3805-403d-9649-11e7090bb789","Type":"ContainerStarted","Data":"e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76"} Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.958342 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" event={"ID":"907ad872-3805-403d-9649-11e7090bb789","Type":"ContainerStarted","Data":"60ad0c1095f6e57afcb208b478242c59b62a66aab37c2e1b514bc9c0fe165fad"} Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.958456 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.960291 4876 generic.go:334] "Generic (PLEG): container finished" podID="b5742635-f4b3-47ee-88be-d59720792b25" containerID="efb2cedad915817582bbb028575ecd4ad526db2eb7c49cf2eb674d1c19766ef6" exitCode=0 Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.960373 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" event={"ID":"b5742635-f4b3-47ee-88be-d59720792b25","Type":"ContainerDied","Data":"efb2cedad915817582bbb028575ecd4ad526db2eb7c49cf2eb674d1c19766ef6"} Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.963641 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" event={"ID":"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd","Type":"ContainerStarted","Data":"1c17eb8586b54f604654478a0606b341378b6a1b6f05ddef61c633dab708d35d"} Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.963678 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" event={"ID":"86e64b2f-3a7e-4221-8f80-6423ccd3e1dd","Type":"ContainerStarted","Data":"019ff4a8dd6ce288dbeb4ef59c16b10ea284c6a87122b464041084c391d3f2ae"} Dec 05 08:18:00 crc kubenswrapper[4876]: I1205 08:18:00.984250 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms85q" event={"ID":"e189296a-a3f0-4937-aa28-c9be917fd59c","Type":"ContainerStarted","Data":"1910c9b8e1870934565e92a285152760d7d6234260c5ffe362cfd7b27c20444d"} Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.011730 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" podStartSLOduration=127.011707567 podStartE2EDuration="2m7.011707567s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:18:01.009805839 +0000 UTC m=+145.498470461" watchObservedRunningTime="2025-12-05 08:18:01.011707567 +0000 UTC m=+145.500372189" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.023413 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-catalog-content\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.023479 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-utilities\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.023534 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgd6s\" (UniqueName: \"kubernetes.io/projected/cb460428-9d76-41ca-bebc-6d1edb40d4b1-kube-api-access-vgd6s\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.036103 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-xb5cw" podStartSLOduration=11.036076248 podStartE2EDuration="11.036076248s" podCreationTimestamp="2025-12-05 08:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:18:01.034696634 +0000 UTC m=+145.523361266" watchObservedRunningTime="2025-12-05 08:18:01.036076248 +0000 UTC m=+145.524740870" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.127551 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-utilities\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.127718 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgd6s\" (UniqueName: \"kubernetes.io/projected/cb460428-9d76-41ca-bebc-6d1edb40d4b1-kube-api-access-vgd6s\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.127990 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-catalog-content\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.128220 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-utilities\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.129957 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-catalog-content\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.156308 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgd6s\" (UniqueName: \"kubernetes.io/projected/cb460428-9d76-41ca-bebc-6d1edb40d4b1-kube-api-access-vgd6s\") pod \"redhat-operators-gfzbx\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.199491 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.255936 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rf92g"] Dec 05 08:18:01 crc kubenswrapper[4876]: W1205 08:18:01.272304 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb48955db_dd1c_4e11_8d01_d465e6e7a75d.slice/crio-85ea3158416128bf6b60af1030d1941902ca7d437b933fef45117b441e9c6028 WatchSource:0}: Error finding container 85ea3158416128bf6b60af1030d1941902ca7d437b933fef45117b441e9c6028: Status 404 returned error can't find the container with id 85ea3158416128bf6b60af1030d1941902ca7d437b933fef45117b441e9c6028 Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.509369 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:01 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:01 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:01 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.509663 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.509878 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gfzbx"] Dec 05 08:18:01 crc kubenswrapper[4876]: W1205 08:18:01.522673 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb460428_9d76_41ca_bebc_6d1edb40d4b1.slice/crio-37249b4a373781882c716349e87c39d9072439577ac7440f3413bee6e2266349 WatchSource:0}: Error finding container 37249b4a373781882c716349e87c39d9072439577ac7440f3413bee6e2266349: Status 404 returned error can't find the container with id 37249b4a373781882c716349e87c39d9072439577ac7440f3413bee6e2266349 Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.834185 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.993473 4876 generic.go:334] "Generic (PLEG): container finished" podID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerID="cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10" exitCode=0 Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.993530 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms85q" event={"ID":"e189296a-a3f0-4937-aa28-c9be917fd59c","Type":"ContainerDied","Data":"cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10"} Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.996612 4876 generic.go:334] "Generic (PLEG): container finished" podID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerID="967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811" exitCode=0 Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.996653 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gfzbx" event={"ID":"cb460428-9d76-41ca-bebc-6d1edb40d4b1","Type":"ContainerDied","Data":"967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811"} Dec 05 08:18:01 crc kubenswrapper[4876]: I1205 08:18:01.996669 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gfzbx" event={"ID":"cb460428-9d76-41ca-bebc-6d1edb40d4b1","Type":"ContainerStarted","Data":"37249b4a373781882c716349e87c39d9072439577ac7440f3413bee6e2266349"} Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.000228 4876 generic.go:334] "Generic (PLEG): container finished" podID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerID="e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99" exitCode=0 Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.000855 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf92g" event={"ID":"b48955db-dd1c-4e11-8d01-d465e6e7a75d","Type":"ContainerDied","Data":"e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99"} Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.000879 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf92g" event={"ID":"b48955db-dd1c-4e11-8d01-d465e6e7a75d","Type":"ContainerStarted","Data":"85ea3158416128bf6b60af1030d1941902ca7d437b933fef45117b441e9c6028"} Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.105773 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vqw7m" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.219552 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.345192 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume\") pod \"b5742635-f4b3-47ee-88be-d59720792b25\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.345291 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf5fb\" (UniqueName: \"kubernetes.io/projected/b5742635-f4b3-47ee-88be-d59720792b25-kube-api-access-lf5fb\") pod \"b5742635-f4b3-47ee-88be-d59720792b25\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.345325 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5742635-f4b3-47ee-88be-d59720792b25-secret-volume\") pod \"b5742635-f4b3-47ee-88be-d59720792b25\" (UID: \"b5742635-f4b3-47ee-88be-d59720792b25\") " Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.346256 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume" (OuterVolumeSpecName: "config-volume") pod "b5742635-f4b3-47ee-88be-d59720792b25" (UID: "b5742635-f4b3-47ee-88be-d59720792b25"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.352859 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5742635-f4b3-47ee-88be-d59720792b25-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b5742635-f4b3-47ee-88be-d59720792b25" (UID: "b5742635-f4b3-47ee-88be-d59720792b25"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.366438 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5742635-f4b3-47ee-88be-d59720792b25-kube-api-access-lf5fb" (OuterVolumeSpecName: "kube-api-access-lf5fb") pod "b5742635-f4b3-47ee-88be-d59720792b25" (UID: "b5742635-f4b3-47ee-88be-d59720792b25"). InnerVolumeSpecName "kube-api-access-lf5fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.447420 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf5fb\" (UniqueName: \"kubernetes.io/projected/b5742635-f4b3-47ee-88be-d59720792b25-kube-api-access-lf5fb\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.447452 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5742635-f4b3-47ee-88be-d59720792b25-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.447461 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5742635-f4b3-47ee-88be-d59720792b25-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.487078 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.488248 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.502773 4876 patch_prober.go:28] interesting pod/apiserver-76f77b778f-zc7rw container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]log ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]etcd ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/generic-apiserver-start-informers ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/max-in-flight-filter ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 05 08:18:02 crc kubenswrapper[4876]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/project.openshift.io-projectcache ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/openshift.io-startinformers ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 05 08:18:02 crc kubenswrapper[4876]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 05 08:18:02 crc kubenswrapper[4876]: livez check failed Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.502839 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" podUID="92e719e9-c64e-4e54-a002-427ebe228761" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.505919 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 08:18:02 crc kubenswrapper[4876]: E1205 08:18:02.511010 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5742635-f4b3-47ee-88be-d59720792b25" containerName="collect-profiles" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.511044 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5742635-f4b3-47ee-88be-d59720792b25" containerName="collect-profiles" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.511815 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:02 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:02 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:02 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.511869 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.511958 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5742635-f4b3-47ee-88be-d59720792b25" containerName="collect-profiles" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.512876 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.516435 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.517090 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.521153 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.521189 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.527936 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.541159 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.649457 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.649544 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.751082 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.751125 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.751222 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.760661 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.760780 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.762507 4876 patch_prober.go:28] interesting pod/console-f9d7485db-plzs7 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.762560 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-plzs7" podUID="75f4a218-73a6-4a91-a8ab-6177ee8d1e45" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.782722 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:02 crc kubenswrapper[4876]: I1205 08:18:02.837977 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.029102 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.032206 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt" event={"ID":"b5742635-f4b3-47ee-88be-d59720792b25","Type":"ContainerDied","Data":"887c6fb4a15e34bd8de607adcbfe43b0635b059582027c82b1ea1ba099163e74"} Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.032274 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="887c6fb4a15e34bd8de607adcbfe43b0635b059582027c82b1ea1ba099163e74" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.036627 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lw86s" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.084769 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 08:18:03 crc kubenswrapper[4876]: W1205 08:18:03.101206 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2e105f35_93bc_43f1_bbb6_4e1cf934e453.slice/crio-75c6e29c609695169f34bfe33801542565722ce3877a0c675fa27d434902be70 WatchSource:0}: Error finding container 75c6e29c609695169f34bfe33801542565722ce3877a0c675fa27d434902be70: Status 404 returned error can't find the container with id 75c6e29c609695169f34bfe33801542565722ce3877a0c675fa27d434902be70 Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.504701 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.508367 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:03 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:03 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:03 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.508432 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.960559 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.974530 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.974587 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.974644 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.974667 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.980408 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.983353 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:18:03 crc kubenswrapper[4876]: I1205 08:18:03.993453 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:18:04 crc kubenswrapper[4876]: I1205 08:18:04.053983 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:18:04 crc kubenswrapper[4876]: I1205 08:18:04.058475 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2e105f35-93bc-43f1-bbb6-4e1cf934e453","Type":"ContainerStarted","Data":"75c6e29c609695169f34bfe33801542565722ce3877a0c675fa27d434902be70"} Dec 05 08:18:04 crc kubenswrapper[4876]: I1205 08:18:04.153665 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:18:04 crc kubenswrapper[4876]: I1205 08:18:04.250583 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 08:18:04 crc kubenswrapper[4876]: W1205 08:18:04.363944 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-1142a11a9be90a382e8f88df7b0a4a4a9ca25e718408f15b54c6d7fb16655a48 WatchSource:0}: Error finding container 1142a11a9be90a382e8f88df7b0a4a4a9ca25e718408f15b54c6d7fb16655a48: Status 404 returned error can't find the container with id 1142a11a9be90a382e8f88df7b0a4a4a9ca25e718408f15b54c6d7fb16655a48 Dec 05 08:18:04 crc kubenswrapper[4876]: I1205 08:18:04.367702 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 08:18:04 crc kubenswrapper[4876]: I1205 08:18:04.512353 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:04 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:04 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:04 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:04 crc kubenswrapper[4876]: I1205 08:18:04.512589 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:04 crc kubenswrapper[4876]: W1205 08:18:04.756178 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-bcf4343436c35833454e2debb87bc462f21a02849414364d3513d0ec725caf22 WatchSource:0}: Error finding container bcf4343436c35833454e2debb87bc462f21a02849414364d3513d0ec725caf22: Status 404 returned error can't find the container with id bcf4343436c35833454e2debb87bc462f21a02849414364d3513d0ec725caf22 Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.069171 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2e105f35-93bc-43f1-bbb6-4e1cf934e453","Type":"ContainerStarted","Data":"d2bce935585650270522a2ec5293457de3010bc34ccd862d455fdc0bbcabfa4c"} Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.073566 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"bcf4343436c35833454e2debb87bc462f21a02849414364d3513d0ec725caf22"} Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.076257 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1d803a3f05fba9584d5dd073aab7e4bd73371b8c57d33b52b01e63e792f182c5"} Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.077664 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4d31bf11403efb652e9d8f2a43a49bffb183b80e327196ba8541d9a5a939c671"} Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.077684 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1142a11a9be90a382e8f88df7b0a4a4a9ca25e718408f15b54c6d7fb16655a48"} Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.078646 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.119194 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.119175873 podStartE2EDuration="3.119175873s" podCreationTimestamp="2025-12-05 08:18:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:18:05.095918689 +0000 UTC m=+149.584583311" watchObservedRunningTime="2025-12-05 08:18:05.119175873 +0000 UTC m=+149.607840495" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.459661 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.460477 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.463744 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.465760 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.465794 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.507566 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:05 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:05 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:05 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.507631 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.604518 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59408954-58c1-496a-bd17-291c32021269-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"59408954-58c1-496a-bd17-291c32021269\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.604739 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59408954-58c1-496a-bd17-291c32021269-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"59408954-58c1-496a-bd17-291c32021269\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.706545 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59408954-58c1-496a-bd17-291c32021269-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"59408954-58c1-496a-bd17-291c32021269\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.706598 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59408954-58c1-496a-bd17-291c32021269-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"59408954-58c1-496a-bd17-291c32021269\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.706688 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59408954-58c1-496a-bd17-291c32021269-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"59408954-58c1-496a-bd17-291c32021269\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.724061 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59408954-58c1-496a-bd17-291c32021269-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"59408954-58c1-496a-bd17-291c32021269\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:05 crc kubenswrapper[4876]: I1205 08:18:05.780673 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:06 crc kubenswrapper[4876]: I1205 08:18:06.088165 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 08:18:06 crc kubenswrapper[4876]: I1205 08:18:06.108325 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"76c192ddf0e627a4ce5cd7e3a827087ab4888635b550e4188ade3664407eb60c"} Dec 05 08:18:06 crc kubenswrapper[4876]: I1205 08:18:06.115994 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7f6debb35caf137833f26220e8c7d010e0d464aad1f497645205099fa89ec1f4"} Dec 05 08:18:06 crc kubenswrapper[4876]: I1205 08:18:06.142786 4876 generic.go:334] "Generic (PLEG): container finished" podID="2e105f35-93bc-43f1-bbb6-4e1cf934e453" containerID="d2bce935585650270522a2ec5293457de3010bc34ccd862d455fdc0bbcabfa4c" exitCode=0 Dec 05 08:18:06 crc kubenswrapper[4876]: I1205 08:18:06.142966 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2e105f35-93bc-43f1-bbb6-4e1cf934e453","Type":"ContainerDied","Data":"d2bce935585650270522a2ec5293457de3010bc34ccd862d455fdc0bbcabfa4c"} Dec 05 08:18:06 crc kubenswrapper[4876]: I1205 08:18:06.508350 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:06 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:06 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:06 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:06 crc kubenswrapper[4876]: I1205 08:18:06.508441 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:07 crc kubenswrapper[4876]: I1205 08:18:07.151351 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"59408954-58c1-496a-bd17-291c32021269","Type":"ContainerStarted","Data":"fbe09ca12814c1f249fd045faebf857437f70deb229a0bd3932710ac03559d6d"} Dec 05 08:18:07 crc kubenswrapper[4876]: I1205 08:18:07.486440 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:18:07 crc kubenswrapper[4876]: I1205 08:18:07.491654 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-zc7rw" Dec 05 08:18:07 crc kubenswrapper[4876]: I1205 08:18:07.507017 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:07 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:07 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:07 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:07 crc kubenswrapper[4876]: I1205 08:18:07.507078 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:08 crc kubenswrapper[4876]: I1205 08:18:08.213654 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:18:08 crc kubenswrapper[4876]: I1205 08:18:08.213965 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:18:08 crc kubenswrapper[4876]: I1205 08:18:08.507533 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:08 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:08 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:08 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:08 crc kubenswrapper[4876]: I1205 08:18:08.507636 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:08 crc kubenswrapper[4876]: I1205 08:18:08.655214 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rd9lr" Dec 05 08:18:09 crc kubenswrapper[4876]: I1205 08:18:09.507476 4876 patch_prober.go:28] interesting pod/router-default-5444994796-6rtbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 08:18:09 crc kubenswrapper[4876]: [-]has-synced failed: reason withheld Dec 05 08:18:09 crc kubenswrapper[4876]: [+]process-running ok Dec 05 08:18:09 crc kubenswrapper[4876]: healthz check failed Dec 05 08:18:09 crc kubenswrapper[4876]: I1205 08:18:09.507572 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-6rtbf" podUID="fed5f23f-513b-449f-adaa-d65ce9beb793" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 08:18:10 crc kubenswrapper[4876]: I1205 08:18:10.506859 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:18:10 crc kubenswrapper[4876]: I1205 08:18:10.509145 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-6rtbf" Dec 05 08:18:12 crc kubenswrapper[4876]: I1205 08:18:12.859966 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:18:12 crc kubenswrapper[4876]: I1205 08:18:12.863983 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:18:13 crc kubenswrapper[4876]: I1205 08:18:13.701003 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:13 crc kubenswrapper[4876]: I1205 08:18:13.852358 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kube-api-access\") pod \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\" (UID: \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\") " Dec 05 08:18:13 crc kubenswrapper[4876]: I1205 08:18:13.852632 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2e105f35-93bc-43f1-bbb6-4e1cf934e453" (UID: "2e105f35-93bc-43f1-bbb6-4e1cf934e453"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:18:13 crc kubenswrapper[4876]: I1205 08:18:13.852726 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kubelet-dir\") pod \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\" (UID: \"2e105f35-93bc-43f1-bbb6-4e1cf934e453\") " Dec 05 08:18:13 crc kubenswrapper[4876]: I1205 08:18:13.854599 4876 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:13 crc kubenswrapper[4876]: I1205 08:18:13.863340 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2e105f35-93bc-43f1-bbb6-4e1cf934e453" (UID: "2e105f35-93bc-43f1-bbb6-4e1cf934e453"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:18:13 crc kubenswrapper[4876]: I1205 08:18:13.955664 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2e105f35-93bc-43f1-bbb6-4e1cf934e453-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:14 crc kubenswrapper[4876]: I1205 08:18:14.207716 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"59408954-58c1-496a-bd17-291c32021269","Type":"ContainerStarted","Data":"93523dca5acc49bd9bf371ddbcaf10d8893222685f4736f2adc81c56af0dda0d"} Dec 05 08:18:14 crc kubenswrapper[4876]: I1205 08:18:14.211402 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2e105f35-93bc-43f1-bbb6-4e1cf934e453","Type":"ContainerDied","Data":"75c6e29c609695169f34bfe33801542565722ce3877a0c675fa27d434902be70"} Dec 05 08:18:14 crc kubenswrapper[4876]: I1205 08:18:14.211431 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75c6e29c609695169f34bfe33801542565722ce3877a0c675fa27d434902be70" Dec 05 08:18:14 crc kubenswrapper[4876]: I1205 08:18:14.211466 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 08:18:14 crc kubenswrapper[4876]: I1205 08:18:14.222995 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=9.222975477 podStartE2EDuration="9.222975477s" podCreationTimestamp="2025-12-05 08:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:18:14.22230342 +0000 UTC m=+158.710968042" watchObservedRunningTime="2025-12-05 08:18:14.222975477 +0000 UTC m=+158.711640089" Dec 05 08:18:16 crc kubenswrapper[4876]: I1205 08:18:16.221571 4876 generic.go:334] "Generic (PLEG): container finished" podID="59408954-58c1-496a-bd17-291c32021269" containerID="93523dca5acc49bd9bf371ddbcaf10d8893222685f4736f2adc81c56af0dda0d" exitCode=0 Dec 05 08:18:16 crc kubenswrapper[4876]: I1205 08:18:16.221612 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"59408954-58c1-496a-bd17-291c32021269","Type":"ContainerDied","Data":"93523dca5acc49bd9bf371ddbcaf10d8893222685f4736f2adc81c56af0dda0d"} Dec 05 08:18:16 crc kubenswrapper[4876]: I1205 08:18:16.326557 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:18:16 crc kubenswrapper[4876]: I1205 08:18:16.341199 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/79225fe3-5793-4316-92da-c227edb92d53-metrics-certs\") pod \"network-metrics-daemon-7wg59\" (UID: \"79225fe3-5793-4316-92da-c227edb92d53\") " pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:18:16 crc kubenswrapper[4876]: I1205 08:18:16.379657 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7wg59" Dec 05 08:18:20 crc kubenswrapper[4876]: I1205 08:18:20.101055 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.013183 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.113373 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59408954-58c1-496a-bd17-291c32021269-kubelet-dir\") pod \"59408954-58c1-496a-bd17-291c32021269\" (UID: \"59408954-58c1-496a-bd17-291c32021269\") " Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.113476 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/59408954-58c1-496a-bd17-291c32021269-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "59408954-58c1-496a-bd17-291c32021269" (UID: "59408954-58c1-496a-bd17-291c32021269"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.113560 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59408954-58c1-496a-bd17-291c32021269-kube-api-access\") pod \"59408954-58c1-496a-bd17-291c32021269\" (UID: \"59408954-58c1-496a-bd17-291c32021269\") " Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.113849 4876 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/59408954-58c1-496a-bd17-291c32021269-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.119556 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59408954-58c1-496a-bd17-291c32021269-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "59408954-58c1-496a-bd17-291c32021269" (UID: "59408954-58c1-496a-bd17-291c32021269"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.214957 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59408954-58c1-496a-bd17-291c32021269-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.267532 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"59408954-58c1-496a-bd17-291c32021269","Type":"ContainerDied","Data":"fbe09ca12814c1f249fd045faebf857437f70deb229a0bd3932710ac03559d6d"} Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.267571 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbe09ca12814c1f249fd045faebf857437f70deb229a0bd3932710ac03559d6d" Dec 05 08:18:23 crc kubenswrapper[4876]: I1205 08:18:23.267629 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 08:18:28 crc kubenswrapper[4876]: E1205 08:18:28.694678 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 08:18:28 crc kubenswrapper[4876]: E1205 08:18:28.695204 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4b7r8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vtmg6_openshift-marketplace(69511d2e-da37-424d-bb34-4791b3292c35): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 08:18:28 crc kubenswrapper[4876]: E1205 08:18:28.696436 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vtmg6" podUID="69511d2e-da37-424d-bb34-4791b3292c35" Dec 05 08:18:28 crc kubenswrapper[4876]: E1205 08:18:28.733950 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 08:18:28 crc kubenswrapper[4876]: E1205 08:18:28.734160 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kdlbc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kdhx2_openshift-marketplace(dfa346f8-9447-4712-b791-21bf718356b3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 08:18:28 crc kubenswrapper[4876]: E1205 08:18:28.735622 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kdhx2" podUID="dfa346f8-9447-4712-b791-21bf718356b3" Dec 05 08:18:30 crc kubenswrapper[4876]: E1205 08:18:30.055056 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vtmg6" podUID="69511d2e-da37-424d-bb34-4791b3292c35" Dec 05 08:18:30 crc kubenswrapper[4876]: E1205 08:18:30.055472 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kdhx2" podUID="dfa346f8-9447-4712-b791-21bf718356b3" Dec 05 08:18:30 crc kubenswrapper[4876]: E1205 08:18:30.133197 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 08:18:30 crc kubenswrapper[4876]: E1205 08:18:30.133364 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zjqqb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-24njp_openshift-marketplace(d0e342ab-27c1-442c-926e-f2e9ed1153c5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 08:18:30 crc kubenswrapper[4876]: E1205 08:18:30.134496 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-24njp" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" Dec 05 08:18:30 crc kubenswrapper[4876]: E1205 08:18:30.154547 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 08:18:30 crc kubenswrapper[4876]: E1205 08:18:30.154735 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mrspc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-nnc6w_openshift-marketplace(17ce8bf7-9544-4849-aa82-c5744a929656): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 08:18:30 crc kubenswrapper[4876]: E1205 08:18:30.155893 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-nnc6w" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" Dec 05 08:18:32 crc kubenswrapper[4876]: E1205 08:18:32.959360 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-nnc6w" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" Dec 05 08:18:32 crc kubenswrapper[4876]: E1205 08:18:32.959441 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-24njp" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" Dec 05 08:18:32 crc kubenswrapper[4876]: E1205 08:18:32.980440 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 08:18:32 crc kubenswrapper[4876]: E1205 08:18:32.980613 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vgd6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gfzbx_openshift-marketplace(cb460428-9d76-41ca-bebc-6d1edb40d4b1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 08:18:32 crc kubenswrapper[4876]: E1205 08:18:32.981801 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gfzbx" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.044132 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.044317 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cnqdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rf92g_openshift-marketplace(b48955db-dd1c-4e11-8d01-d465e6e7a75d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.045478 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rf92g" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" Dec 05 08:18:33 crc kubenswrapper[4876]: I1205 08:18:33.310801 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-tvfkn" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.913634 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rf92g" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.967562 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.967961 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cxm4w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ms85q_openshift-marketplace(e189296a-a3f0-4937-aa28-c9be917fd59c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.972759 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ms85q" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.982670 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.982775 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zt7qv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4g9z7_openshift-marketplace(5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 08:18:33 crc kubenswrapper[4876]: E1205 08:18:33.986917 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4g9z7" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" Dec 05 08:18:34 crc kubenswrapper[4876]: I1205 08:18:34.062007 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 08:18:34 crc kubenswrapper[4876]: I1205 08:18:34.299141 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7wg59"] Dec 05 08:18:34 crc kubenswrapper[4876]: I1205 08:18:34.324871 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7wg59" event={"ID":"79225fe3-5793-4316-92da-c227edb92d53","Type":"ContainerStarted","Data":"a3acc6240756b9c6bfd443b56dfa548fe7ddf03ef38bb184858b83c2aff6252c"} Dec 05 08:18:34 crc kubenswrapper[4876]: E1205 08:18:34.338891 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ms85q" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" Dec 05 08:18:34 crc kubenswrapper[4876]: E1205 08:18:34.339554 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4g9z7" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" Dec 05 08:18:35 crc kubenswrapper[4876]: I1205 08:18:35.333122 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7wg59" event={"ID":"79225fe3-5793-4316-92da-c227edb92d53","Type":"ContainerStarted","Data":"525906e91589c3d8de1929692def5b2ea509a6cb8d88421f1b3a8279707ef5cf"} Dec 05 08:18:35 crc kubenswrapper[4876]: I1205 08:18:35.333416 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7wg59" event={"ID":"79225fe3-5793-4316-92da-c227edb92d53","Type":"ContainerStarted","Data":"6d8381c0ae670b35bc89601596420840a0a4d1c861b04fe69524c0bf936f19bd"} Dec 05 08:18:35 crc kubenswrapper[4876]: I1205 08:18:35.354529 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7wg59" podStartSLOduration=161.354507948 podStartE2EDuration="2m41.354507948s" podCreationTimestamp="2025-12-05 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:18:35.347674346 +0000 UTC m=+179.836339008" watchObservedRunningTime="2025-12-05 08:18:35.354507948 +0000 UTC m=+179.843172580" Dec 05 08:18:38 crc kubenswrapper[4876]: I1205 08:18:38.213762 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:18:38 crc kubenswrapper[4876]: I1205 08:18:38.214372 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.660876 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 08:18:39 crc kubenswrapper[4876]: E1205 08:18:39.661574 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e105f35-93bc-43f1-bbb6-4e1cf934e453" containerName="pruner" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.661595 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e105f35-93bc-43f1-bbb6-4e1cf934e453" containerName="pruner" Dec 05 08:18:39 crc kubenswrapper[4876]: E1205 08:18:39.661634 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59408954-58c1-496a-bd17-291c32021269" containerName="pruner" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.661642 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="59408954-58c1-496a-bd17-291c32021269" containerName="pruner" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.661757 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e105f35-93bc-43f1-bbb6-4e1cf934e453" containerName="pruner" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.661767 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="59408954-58c1-496a-bd17-291c32021269" containerName="pruner" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.662163 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.664103 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.664451 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.673424 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.820289 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.820424 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.921571 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.921657 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.921754 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.941433 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:39 crc kubenswrapper[4876]: I1205 08:18:39.991661 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:40 crc kubenswrapper[4876]: I1205 08:18:40.406861 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 08:18:40 crc kubenswrapper[4876]: W1205 08:18:40.416497 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod81af2eb0_1a4f_4f20_8753_16cec5cc9e95.slice/crio-11e4a26ff8791fc61348482fa8dd5090803a6f3f3988bf0a6aa3f7636e6bf5f6 WatchSource:0}: Error finding container 11e4a26ff8791fc61348482fa8dd5090803a6f3f3988bf0a6aa3f7636e6bf5f6: Status 404 returned error can't find the container with id 11e4a26ff8791fc61348482fa8dd5090803a6f3f3988bf0a6aa3f7636e6bf5f6 Dec 05 08:18:41 crc kubenswrapper[4876]: I1205 08:18:41.364054 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81af2eb0-1a4f-4f20-8753-16cec5cc9e95","Type":"ContainerStarted","Data":"ba32ffc92ebad9794ecbbe013587a3ef636539ac2c474773ca4b588ca441f359"} Dec 05 08:18:41 crc kubenswrapper[4876]: I1205 08:18:41.364435 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81af2eb0-1a4f-4f20-8753-16cec5cc9e95","Type":"ContainerStarted","Data":"11e4a26ff8791fc61348482fa8dd5090803a6f3f3988bf0a6aa3f7636e6bf5f6"} Dec 05 08:18:41 crc kubenswrapper[4876]: I1205 08:18:41.376846 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.376830441 podStartE2EDuration="2.376830441s" podCreationTimestamp="2025-12-05 08:18:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:18:41.375349104 +0000 UTC m=+185.864013736" watchObservedRunningTime="2025-12-05 08:18:41.376830441 +0000 UTC m=+185.865495063" Dec 05 08:18:42 crc kubenswrapper[4876]: I1205 08:18:42.370154 4876 generic.go:334] "Generic (PLEG): container finished" podID="81af2eb0-1a4f-4f20-8753-16cec5cc9e95" containerID="ba32ffc92ebad9794ecbbe013587a3ef636539ac2c474773ca4b588ca441f359" exitCode=0 Dec 05 08:18:42 crc kubenswrapper[4876]: I1205 08:18:42.370432 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81af2eb0-1a4f-4f20-8753-16cec5cc9e95","Type":"ContainerDied","Data":"ba32ffc92ebad9794ecbbe013587a3ef636539ac2c474773ca4b588ca441f359"} Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.376762 4876 generic.go:334] "Generic (PLEG): container finished" podID="69511d2e-da37-424d-bb34-4791b3292c35" containerID="94d777af08a96b450c3f4189af6a7e710a9875cba82a3893125592a897b420c5" exitCode=0 Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.376840 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtmg6" event={"ID":"69511d2e-da37-424d-bb34-4791b3292c35","Type":"ContainerDied","Data":"94d777af08a96b450c3f4189af6a7e710a9875cba82a3893125592a897b420c5"} Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.612017 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.770745 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kube-api-access\") pod \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\" (UID: \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\") " Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.770843 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kubelet-dir\") pod \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\" (UID: \"81af2eb0-1a4f-4f20-8753-16cec5cc9e95\") " Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.771039 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "81af2eb0-1a4f-4f20-8753-16cec5cc9e95" (UID: "81af2eb0-1a4f-4f20-8753-16cec5cc9e95"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.777312 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "81af2eb0-1a4f-4f20-8753-16cec5cc9e95" (UID: "81af2eb0-1a4f-4f20-8753-16cec5cc9e95"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.872442 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:43 crc kubenswrapper[4876]: I1205 08:18:43.872477 4876 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/81af2eb0-1a4f-4f20-8753-16cec5cc9e95-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:44 crc kubenswrapper[4876]: I1205 08:18:44.391065 4876 generic.go:334] "Generic (PLEG): container finished" podID="dfa346f8-9447-4712-b791-21bf718356b3" containerID="352c57fdbbfb1e20bdb97e12dd2aae7c1ae76d1c28e3f24cc34dee9e20bf0dd5" exitCode=0 Dec 05 08:18:44 crc kubenswrapper[4876]: I1205 08:18:44.391101 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdhx2" event={"ID":"dfa346f8-9447-4712-b791-21bf718356b3","Type":"ContainerDied","Data":"352c57fdbbfb1e20bdb97e12dd2aae7c1ae76d1c28e3f24cc34dee9e20bf0dd5"} Dec 05 08:18:44 crc kubenswrapper[4876]: I1205 08:18:44.396359 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtmg6" event={"ID":"69511d2e-da37-424d-bb34-4791b3292c35","Type":"ContainerStarted","Data":"a59b0d6fff5cfe773dc8af339b59e7d6ea1fa88b5dd56460df073d7b0e5559bd"} Dec 05 08:18:44 crc kubenswrapper[4876]: I1205 08:18:44.397786 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"81af2eb0-1a4f-4f20-8753-16cec5cc9e95","Type":"ContainerDied","Data":"11e4a26ff8791fc61348482fa8dd5090803a6f3f3988bf0a6aa3f7636e6bf5f6"} Dec 05 08:18:44 crc kubenswrapper[4876]: I1205 08:18:44.397812 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11e4a26ff8791fc61348482fa8dd5090803a6f3f3988bf0a6aa3f7636e6bf5f6" Dec 05 08:18:44 crc kubenswrapper[4876]: I1205 08:18:44.397848 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 08:18:44 crc kubenswrapper[4876]: I1205 08:18:44.425593 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vtmg6" podStartSLOduration=3.541348889 podStartE2EDuration="47.425579261s" podCreationTimestamp="2025-12-05 08:17:57 +0000 UTC" firstStartedPulling="2025-12-05 08:17:59.903550759 +0000 UTC m=+144.392215381" lastFinishedPulling="2025-12-05 08:18:43.787781131 +0000 UTC m=+188.276445753" observedRunningTime="2025-12-05 08:18:44.424647527 +0000 UTC m=+188.913312149" watchObservedRunningTime="2025-12-05 08:18:44.425579261 +0000 UTC m=+188.914243883" Dec 05 08:18:45 crc kubenswrapper[4876]: I1205 08:18:45.403298 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnc6w" event={"ID":"17ce8bf7-9544-4849-aa82-c5744a929656","Type":"ContainerStarted","Data":"93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2"} Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.413878 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdhx2" event={"ID":"dfa346f8-9447-4712-b791-21bf718356b3","Type":"ContainerStarted","Data":"a4a5bd99215d91c38e65ded9d389c4f34f800fef598b9d0019f449c6d2362c5e"} Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.415667 4876 generic.go:334] "Generic (PLEG): container finished" podID="17ce8bf7-9544-4849-aa82-c5744a929656" containerID="93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2" exitCode=0 Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.415689 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnc6w" event={"ID":"17ce8bf7-9544-4849-aa82-c5744a929656","Type":"ContainerDied","Data":"93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2"} Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.433338 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kdhx2" podStartSLOduration=3.8396198569999997 podStartE2EDuration="49.433320429s" podCreationTimestamp="2025-12-05 08:17:57 +0000 UTC" firstStartedPulling="2025-12-05 08:17:59.81751876 +0000 UTC m=+144.306183382" lastFinishedPulling="2025-12-05 08:18:45.411219332 +0000 UTC m=+189.899883954" observedRunningTime="2025-12-05 08:18:46.430117028 +0000 UTC m=+190.918781650" watchObservedRunningTime="2025-12-05 08:18:46.433320429 +0000 UTC m=+190.921985051" Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.851847 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 08:18:46 crc kubenswrapper[4876]: E1205 08:18:46.852081 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81af2eb0-1a4f-4f20-8753-16cec5cc9e95" containerName="pruner" Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.852099 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="81af2eb0-1a4f-4f20-8753-16cec5cc9e95" containerName="pruner" Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.852216 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="81af2eb0-1a4f-4f20-8753-16cec5cc9e95" containerName="pruner" Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.852610 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.854407 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.854695 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 08:18:46 crc kubenswrapper[4876]: I1205 08:18:46.865080 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.013612 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-var-lock\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.013715 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kube-api-access\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.013794 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.115550 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.115656 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-var-lock\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.115682 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kube-api-access\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.116128 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.116176 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-var-lock\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.137821 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kube-api-access\") pod \"installer-9-crc\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.172830 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.421445 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gfzbx" event={"ID":"cb460428-9d76-41ca-bebc-6d1edb40d4b1","Type":"ContainerStarted","Data":"d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615"} Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.423055 4876 generic.go:334] "Generic (PLEG): container finished" podID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerID="c29191fa8082a0feeb327d7a61752e9780b5b52720eb488a47f3423bbcf79104" exitCode=0 Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.423099 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24njp" event={"ID":"d0e342ab-27c1-442c-926e-f2e9ed1153c5","Type":"ContainerDied","Data":"c29191fa8082a0feeb327d7a61752e9780b5b52720eb488a47f3423bbcf79104"} Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.425739 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnc6w" event={"ID":"17ce8bf7-9544-4849-aa82-c5744a929656","Type":"ContainerStarted","Data":"750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142"} Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.471144 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nnc6w" podStartSLOduration=3.228205449 podStartE2EDuration="50.47112588s" podCreationTimestamp="2025-12-05 08:17:57 +0000 UTC" firstStartedPulling="2025-12-05 08:17:59.817710775 +0000 UTC m=+144.306375397" lastFinishedPulling="2025-12-05 08:18:47.060631206 +0000 UTC m=+191.549295828" observedRunningTime="2025-12-05 08:18:47.470682519 +0000 UTC m=+191.959347151" watchObservedRunningTime="2025-12-05 08:18:47.47112588 +0000 UTC m=+191.959790502" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.563668 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.563708 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.567227 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.620629 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.860432 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:18:47 crc kubenswrapper[4876]: I1205 08:18:47.860829 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.101234 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.101529 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.140880 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.431725 4876 generic.go:334] "Generic (PLEG): container finished" podID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerID="096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e" exitCode=0 Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.431794 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf92g" event={"ID":"b48955db-dd1c-4e11-8d01-d465e6e7a75d","Type":"ContainerDied","Data":"096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e"} Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.434150 4876 generic.go:334] "Generic (PLEG): container finished" podID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerID="a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7" exitCode=0 Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.434234 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms85q" event={"ID":"e189296a-a3f0-4937-aa28-c9be917fd59c","Type":"ContainerDied","Data":"a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7"} Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.438491 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d67527e2-6101-4cad-9a5e-3a5a0feed0b9","Type":"ContainerStarted","Data":"bff97825b0319fe15bfd495f36ceb6576c3f3476756c6a6752109aaf85816a05"} Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.438528 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d67527e2-6101-4cad-9a5e-3a5a0feed0b9","Type":"ContainerStarted","Data":"03139afe014bf515fd6c3137c8a5268f25b0b0ced1a68f2d86705e8f44f2273c"} Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.440104 4876 generic.go:334] "Generic (PLEG): container finished" podID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerID="d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615" exitCode=0 Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.440219 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gfzbx" event={"ID":"cb460428-9d76-41ca-bebc-6d1edb40d4b1","Type":"ContainerDied","Data":"d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615"} Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.481296 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.506834 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.506816769 podStartE2EDuration="2.506816769s" podCreationTimestamp="2025-12-05 08:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:18:48.505722371 +0000 UTC m=+192.994386993" watchObservedRunningTime="2025-12-05 08:18:48.506816769 +0000 UTC m=+192.995481391" Dec 05 08:18:48 crc kubenswrapper[4876]: I1205 08:18:48.898490 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-nnc6w" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="registry-server" probeResult="failure" output=< Dec 05 08:18:48 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 08:18:48 crc kubenswrapper[4876]: > Dec 05 08:18:49 crc kubenswrapper[4876]: I1205 08:18:49.451770 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24njp" event={"ID":"d0e342ab-27c1-442c-926e-f2e9ed1153c5","Type":"ContainerStarted","Data":"6829704f713e23f35a806d8d08caaa6dbc5d15d16385b32192abb794d572e49a"} Dec 05 08:18:49 crc kubenswrapper[4876]: I1205 08:18:49.472832 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-24njp" podStartSLOduration=3.218607408 podStartE2EDuration="52.472813306s" podCreationTimestamp="2025-12-05 08:17:57 +0000 UTC" firstStartedPulling="2025-12-05 08:17:59.861152755 +0000 UTC m=+144.349817377" lastFinishedPulling="2025-12-05 08:18:49.115358633 +0000 UTC m=+193.604023275" observedRunningTime="2025-12-05 08:18:49.470015786 +0000 UTC m=+193.958680448" watchObservedRunningTime="2025-12-05 08:18:49.472813306 +0000 UTC m=+193.961477918" Dec 05 08:18:49 crc kubenswrapper[4876]: I1205 08:18:49.855775 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vtmg6"] Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.459369 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms85q" event={"ID":"e189296a-a3f0-4937-aa28-c9be917fd59c","Type":"ContainerStarted","Data":"977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371"} Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.462409 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gfzbx" event={"ID":"cb460428-9d76-41ca-bebc-6d1edb40d4b1","Type":"ContainerStarted","Data":"fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4"} Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.464553 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf92g" event={"ID":"b48955db-dd1c-4e11-8d01-d465e6e7a75d","Type":"ContainerStarted","Data":"f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5"} Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.467217 4876 generic.go:334] "Generic (PLEG): container finished" podID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerID="a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533" exitCode=0 Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.467296 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4g9z7" event={"ID":"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e","Type":"ContainerDied","Data":"a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533"} Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.540674 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gfzbx" podStartSLOduration=4.043422529 podStartE2EDuration="50.540659832s" podCreationTimestamp="2025-12-05 08:18:00 +0000 UTC" firstStartedPulling="2025-12-05 08:18:03.036057002 +0000 UTC m=+147.524721624" lastFinishedPulling="2025-12-05 08:18:49.533294265 +0000 UTC m=+194.021958927" observedRunningTime="2025-12-05 08:18:50.539266187 +0000 UTC m=+195.027930809" watchObservedRunningTime="2025-12-05 08:18:50.540659832 +0000 UTC m=+195.029324454" Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.542678 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ms85q" podStartSLOduration=4.100836699 podStartE2EDuration="51.542669432s" podCreationTimestamp="2025-12-05 08:17:59 +0000 UTC" firstStartedPulling="2025-12-05 08:18:01.995487842 +0000 UTC m=+146.484152474" lastFinishedPulling="2025-12-05 08:18:49.437320585 +0000 UTC m=+193.925985207" observedRunningTime="2025-12-05 08:18:50.516216408 +0000 UTC m=+195.004881030" watchObservedRunningTime="2025-12-05 08:18:50.542669432 +0000 UTC m=+195.031334054" Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.574423 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rf92g" podStartSLOduration=3.096373557 podStartE2EDuration="50.574403089s" podCreationTimestamp="2025-12-05 08:18:00 +0000 UTC" firstStartedPulling="2025-12-05 08:18:02.001624306 +0000 UTC m=+146.490288928" lastFinishedPulling="2025-12-05 08:18:49.479653838 +0000 UTC m=+193.968318460" observedRunningTime="2025-12-05 08:18:50.57086355 +0000 UTC m=+195.059528172" watchObservedRunningTime="2025-12-05 08:18:50.574403089 +0000 UTC m=+195.063067711" Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.773690 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:50 crc kubenswrapper[4876]: I1205 08:18:50.773822 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:18:51 crc kubenswrapper[4876]: I1205 08:18:51.200550 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:51 crc kubenswrapper[4876]: I1205 08:18:51.200886 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:18:51 crc kubenswrapper[4876]: I1205 08:18:51.471754 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vtmg6" podUID="69511d2e-da37-424d-bb34-4791b3292c35" containerName="registry-server" containerID="cri-o://a59b0d6fff5cfe773dc8af339b59e7d6ea1fa88b5dd56460df073d7b0e5559bd" gracePeriod=2 Dec 05 08:18:51 crc kubenswrapper[4876]: I1205 08:18:51.834320 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rf92g" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="registry-server" probeResult="failure" output=< Dec 05 08:18:51 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 08:18:51 crc kubenswrapper[4876]: > Dec 05 08:18:52 crc kubenswrapper[4876]: I1205 08:18:52.243338 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gfzbx" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="registry-server" probeResult="failure" output=< Dec 05 08:18:52 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 08:18:52 crc kubenswrapper[4876]: > Dec 05 08:18:53 crc kubenswrapper[4876]: I1205 08:18:53.482892 4876 generic.go:334] "Generic (PLEG): container finished" podID="69511d2e-da37-424d-bb34-4791b3292c35" containerID="a59b0d6fff5cfe773dc8af339b59e7d6ea1fa88b5dd56460df073d7b0e5559bd" exitCode=0 Dec 05 08:18:53 crc kubenswrapper[4876]: I1205 08:18:53.482933 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtmg6" event={"ID":"69511d2e-da37-424d-bb34-4791b3292c35","Type":"ContainerDied","Data":"a59b0d6fff5cfe773dc8af339b59e7d6ea1fa88b5dd56460df073d7b0e5559bd"} Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.494596 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtmg6" event={"ID":"69511d2e-da37-424d-bb34-4791b3292c35","Type":"ContainerDied","Data":"58a9a94da32b6650d9c03b97e6073b8547847ae02d9c03a01881880ba58bdcf3"} Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.494913 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58a9a94da32b6650d9c03b97e6073b8547847ae02d9c03a01881880ba58bdcf3" Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.527766 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.709102 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-utilities\") pod \"69511d2e-da37-424d-bb34-4791b3292c35\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.709205 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b7r8\" (UniqueName: \"kubernetes.io/projected/69511d2e-da37-424d-bb34-4791b3292c35-kube-api-access-4b7r8\") pod \"69511d2e-da37-424d-bb34-4791b3292c35\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.709229 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-catalog-content\") pod \"69511d2e-da37-424d-bb34-4791b3292c35\" (UID: \"69511d2e-da37-424d-bb34-4791b3292c35\") " Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.710007 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-utilities" (OuterVolumeSpecName: "utilities") pod "69511d2e-da37-424d-bb34-4791b3292c35" (UID: "69511d2e-da37-424d-bb34-4791b3292c35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.719717 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69511d2e-da37-424d-bb34-4791b3292c35-kube-api-access-4b7r8" (OuterVolumeSpecName: "kube-api-access-4b7r8") pod "69511d2e-da37-424d-bb34-4791b3292c35" (UID: "69511d2e-da37-424d-bb34-4791b3292c35"). InnerVolumeSpecName "kube-api-access-4b7r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.766305 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69511d2e-da37-424d-bb34-4791b3292c35" (UID: "69511d2e-da37-424d-bb34-4791b3292c35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.811078 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b7r8\" (UniqueName: \"kubernetes.io/projected/69511d2e-da37-424d-bb34-4791b3292c35-kube-api-access-4b7r8\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.811117 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:54 crc kubenswrapper[4876]: I1205 08:18:54.811134 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69511d2e-da37-424d-bb34-4791b3292c35-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:18:55 crc kubenswrapper[4876]: I1205 08:18:55.500580 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtmg6" Dec 05 08:18:55 crc kubenswrapper[4876]: I1205 08:18:55.540771 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vtmg6"] Dec 05 08:18:55 crc kubenswrapper[4876]: I1205 08:18:55.544980 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vtmg6"] Dec 05 08:18:55 crc kubenswrapper[4876]: I1205 08:18:55.832431 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69511d2e-da37-424d-bb34-4791b3292c35" path="/var/lib/kubelet/pods/69511d2e-da37-424d-bb34-4791b3292c35/volumes" Dec 05 08:18:57 crc kubenswrapper[4876]: I1205 08:18:57.603505 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:18:57 crc kubenswrapper[4876]: I1205 08:18:57.904956 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:18:57 crc kubenswrapper[4876]: I1205 08:18:57.945171 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:18:58 crc kubenswrapper[4876]: I1205 08:18:58.252027 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:18:58 crc kubenswrapper[4876]: I1205 08:18:58.252377 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:18:58 crc kubenswrapper[4876]: I1205 08:18:58.315263 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:18:58 crc kubenswrapper[4876]: I1205 08:18:58.522756 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4g9z7" event={"ID":"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e","Type":"ContainerStarted","Data":"f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c"} Dec 05 08:18:58 crc kubenswrapper[4876]: I1205 08:18:58.566737 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:18:59 crc kubenswrapper[4876]: I1205 08:18:59.549367 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4g9z7" podStartSLOduration=4.071440793 podStartE2EDuration="1m0.549348739s" podCreationTimestamp="2025-12-05 08:17:59 +0000 UTC" firstStartedPulling="2025-12-05 08:18:00.954566182 +0000 UTC m=+145.443230804" lastFinishedPulling="2025-12-05 08:18:57.432474108 +0000 UTC m=+201.921138750" observedRunningTime="2025-12-05 08:18:59.548484636 +0000 UTC m=+204.037149258" watchObservedRunningTime="2025-12-05 08:18:59.549348739 +0000 UTC m=+204.038013381" Dec 05 08:18:59 crc kubenswrapper[4876]: I1205 08:18:59.727467 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:18:59 crc kubenswrapper[4876]: I1205 08:18:59.727516 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:18:59 crc kubenswrapper[4876]: I1205 08:18:59.781319 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:19:00 crc kubenswrapper[4876]: I1205 08:19:00.057600 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24njp"] Dec 05 08:19:00 crc kubenswrapper[4876]: I1205 08:19:00.168677 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:19:00 crc kubenswrapper[4876]: I1205 08:19:00.168999 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:19:00 crc kubenswrapper[4876]: I1205 08:19:00.214876 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:19:00 crc kubenswrapper[4876]: I1205 08:19:00.534554 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-24njp" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerName="registry-server" containerID="cri-o://6829704f713e23f35a806d8d08caaa6dbc5d15d16385b32192abb794d572e49a" gracePeriod=2 Dec 05 08:19:00 crc kubenswrapper[4876]: I1205 08:19:00.585391 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:19:00 crc kubenswrapper[4876]: I1205 08:19:00.823460 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:19:00 crc kubenswrapper[4876]: I1205 08:19:00.874883 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:19:01 crc kubenswrapper[4876]: I1205 08:19:01.263110 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:19:01 crc kubenswrapper[4876]: I1205 08:19:01.325015 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:19:02 crc kubenswrapper[4876]: I1205 08:19:02.460756 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ms85q"] Dec 05 08:19:02 crc kubenswrapper[4876]: I1205 08:19:02.548434 4876 generic.go:334] "Generic (PLEG): container finished" podID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerID="6829704f713e23f35a806d8d08caaa6dbc5d15d16385b32192abb794d572e49a" exitCode=0 Dec 05 08:19:02 crc kubenswrapper[4876]: I1205 08:19:02.549338 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24njp" event={"ID":"d0e342ab-27c1-442c-926e-f2e9ed1153c5","Type":"ContainerDied","Data":"6829704f713e23f35a806d8d08caaa6dbc5d15d16385b32192abb794d572e49a"} Dec 05 08:19:02 crc kubenswrapper[4876]: I1205 08:19:02.845034 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:19:02 crc kubenswrapper[4876]: I1205 08:19:02.922126 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjqqb\" (UniqueName: \"kubernetes.io/projected/d0e342ab-27c1-442c-926e-f2e9ed1153c5-kube-api-access-zjqqb\") pod \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " Dec 05 08:19:02 crc kubenswrapper[4876]: I1205 08:19:02.922186 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-catalog-content\") pod \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " Dec 05 08:19:02 crc kubenswrapper[4876]: I1205 08:19:02.929548 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0e342ab-27c1-442c-926e-f2e9ed1153c5-kube-api-access-zjqqb" (OuterVolumeSpecName: "kube-api-access-zjqqb") pod "d0e342ab-27c1-442c-926e-f2e9ed1153c5" (UID: "d0e342ab-27c1-442c-926e-f2e9ed1153c5"). InnerVolumeSpecName "kube-api-access-zjqqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:19:02 crc kubenswrapper[4876]: I1205 08:19:02.984035 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0e342ab-27c1-442c-926e-f2e9ed1153c5" (UID: "d0e342ab-27c1-442c-926e-f2e9ed1153c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.023359 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-utilities\") pod \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\" (UID: \"d0e342ab-27c1-442c-926e-f2e9ed1153c5\") " Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.024651 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjqqb\" (UniqueName: \"kubernetes.io/projected/d0e342ab-27c1-442c-926e-f2e9ed1153c5-kube-api-access-zjqqb\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.024834 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.025276 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-utilities" (OuterVolumeSpecName: "utilities") pod "d0e342ab-27c1-442c-926e-f2e9ed1153c5" (UID: "d0e342ab-27c1-442c-926e-f2e9ed1153c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.126261 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e342ab-27c1-442c-926e-f2e9ed1153c5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.562793 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24njp" event={"ID":"d0e342ab-27c1-442c-926e-f2e9ed1153c5","Type":"ContainerDied","Data":"5df63ae6d04ec33016709bede1bbc2fdf09b6bcc57cc913938a1da88ce62f130"} Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.562818 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24njp" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.563096 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ms85q" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerName="registry-server" containerID="cri-o://977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371" gracePeriod=2 Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.563281 4876 scope.go:117] "RemoveContainer" containerID="6829704f713e23f35a806d8d08caaa6dbc5d15d16385b32192abb794d572e49a" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.592285 4876 scope.go:117] "RemoveContainer" containerID="c29191fa8082a0feeb327d7a61752e9780b5b52720eb488a47f3423bbcf79104" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.611717 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24njp"] Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.617524 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-24njp"] Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.645467 4876 scope.go:117] "RemoveContainer" containerID="4bfe52d0ca7cb917d415116724ed0c1ea0d94c931056e842d45b02e67eca05e6" Dec 05 08:19:03 crc kubenswrapper[4876]: I1205 08:19:03.837268 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" path="/var/lib/kubelet/pods/d0e342ab-27c1-442c-926e-f2e9ed1153c5/volumes" Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.858282 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gfzbx"] Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.859140 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gfzbx" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="registry-server" containerID="cri-o://fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4" gracePeriod=2 Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.878260 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.949876 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-utilities\") pod \"e189296a-a3f0-4937-aa28-c9be917fd59c\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.949928 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-catalog-content\") pod \"e189296a-a3f0-4937-aa28-c9be917fd59c\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.949999 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxm4w\" (UniqueName: \"kubernetes.io/projected/e189296a-a3f0-4937-aa28-c9be917fd59c-kube-api-access-cxm4w\") pod \"e189296a-a3f0-4937-aa28-c9be917fd59c\" (UID: \"e189296a-a3f0-4937-aa28-c9be917fd59c\") " Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.951199 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-utilities" (OuterVolumeSpecName: "utilities") pod "e189296a-a3f0-4937-aa28-c9be917fd59c" (UID: "e189296a-a3f0-4937-aa28-c9be917fd59c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.961052 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e189296a-a3f0-4937-aa28-c9be917fd59c-kube-api-access-cxm4w" (OuterVolumeSpecName: "kube-api-access-cxm4w") pod "e189296a-a3f0-4937-aa28-c9be917fd59c" (UID: "e189296a-a3f0-4937-aa28-c9be917fd59c"). InnerVolumeSpecName "kube-api-access-cxm4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:19:04 crc kubenswrapper[4876]: I1205 08:19:04.974881 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e189296a-a3f0-4937-aa28-c9be917fd59c" (UID: "e189296a-a3f0-4937-aa28-c9be917fd59c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.051251 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxm4w\" (UniqueName: \"kubernetes.io/projected/e189296a-a3f0-4937-aa28-c9be917fd59c-kube-api-access-cxm4w\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.051305 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.051328 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e189296a-a3f0-4937-aa28-c9be917fd59c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.195477 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.355570 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-catalog-content\") pod \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.355687 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-utilities\") pod \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.355768 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgd6s\" (UniqueName: \"kubernetes.io/projected/cb460428-9d76-41ca-bebc-6d1edb40d4b1-kube-api-access-vgd6s\") pod \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\" (UID: \"cb460428-9d76-41ca-bebc-6d1edb40d4b1\") " Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.357389 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-utilities" (OuterVolumeSpecName: "utilities") pod "cb460428-9d76-41ca-bebc-6d1edb40d4b1" (UID: "cb460428-9d76-41ca-bebc-6d1edb40d4b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.362575 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb460428-9d76-41ca-bebc-6d1edb40d4b1-kube-api-access-vgd6s" (OuterVolumeSpecName: "kube-api-access-vgd6s") pod "cb460428-9d76-41ca-bebc-6d1edb40d4b1" (UID: "cb460428-9d76-41ca-bebc-6d1edb40d4b1"). InnerVolumeSpecName "kube-api-access-vgd6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.459849 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.459953 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgd6s\" (UniqueName: \"kubernetes.io/projected/cb460428-9d76-41ca-bebc-6d1edb40d4b1-kube-api-access-vgd6s\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.548696 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb460428-9d76-41ca-bebc-6d1edb40d4b1" (UID: "cb460428-9d76-41ca-bebc-6d1edb40d4b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.561152 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb460428-9d76-41ca-bebc-6d1edb40d4b1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.577959 4876 generic.go:334] "Generic (PLEG): container finished" podID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerID="fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4" exitCode=0 Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.578058 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gfzbx" event={"ID":"cb460428-9d76-41ca-bebc-6d1edb40d4b1","Type":"ContainerDied","Data":"fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4"} Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.578146 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gfzbx" event={"ID":"cb460428-9d76-41ca-bebc-6d1edb40d4b1","Type":"ContainerDied","Data":"37249b4a373781882c716349e87c39d9072439577ac7440f3413bee6e2266349"} Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.578186 4876 scope.go:117] "RemoveContainer" containerID="fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.578373 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gfzbx" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.594171 4876 generic.go:334] "Generic (PLEG): container finished" podID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerID="977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371" exitCode=0 Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.594270 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms85q" event={"ID":"e189296a-a3f0-4937-aa28-c9be917fd59c","Type":"ContainerDied","Data":"977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371"} Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.594310 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ms85q" event={"ID":"e189296a-a3f0-4937-aa28-c9be917fd59c","Type":"ContainerDied","Data":"1910c9b8e1870934565e92a285152760d7d6234260c5ffe362cfd7b27c20444d"} Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.594428 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ms85q" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.623605 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gfzbx"] Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.625602 4876 scope.go:117] "RemoveContainer" containerID="d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.632179 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gfzbx"] Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.647720 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ms85q"] Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.655752 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ms85q"] Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.657683 4876 scope.go:117] "RemoveContainer" containerID="967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.676019 4876 scope.go:117] "RemoveContainer" containerID="fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4" Dec 05 08:19:05 crc kubenswrapper[4876]: E1205 08:19:05.676588 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4\": container with ID starting with fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4 not found: ID does not exist" containerID="fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.676665 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4"} err="failed to get container status \"fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4\": rpc error: code = NotFound desc = could not find container \"fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4\": container with ID starting with fb34149f2011e6c4df4f099eb2ef2789dbbcd1aeabca4556c43881fbb5d695f4 not found: ID does not exist" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.676733 4876 scope.go:117] "RemoveContainer" containerID="d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615" Dec 05 08:19:05 crc kubenswrapper[4876]: E1205 08:19:05.677269 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615\": container with ID starting with d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615 not found: ID does not exist" containerID="d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.677334 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615"} err="failed to get container status \"d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615\": rpc error: code = NotFound desc = could not find container \"d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615\": container with ID starting with d6596c210166ad2eb1c57a359f6cdc0e6fa054beab10ed2aa702211ac23da615 not found: ID does not exist" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.677376 4876 scope.go:117] "RemoveContainer" containerID="967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811" Dec 05 08:19:05 crc kubenswrapper[4876]: E1205 08:19:05.677845 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811\": container with ID starting with 967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811 not found: ID does not exist" containerID="967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.677971 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811"} err="failed to get container status \"967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811\": rpc error: code = NotFound desc = could not find container \"967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811\": container with ID starting with 967c3f2071ae9088a1c4b042a4442e71620f528242d9af61dcd4e0daaa5f3811 not found: ID does not exist" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.678010 4876 scope.go:117] "RemoveContainer" containerID="977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.695991 4876 scope.go:117] "RemoveContainer" containerID="a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.713353 4876 scope.go:117] "RemoveContainer" containerID="cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.736777 4876 scope.go:117] "RemoveContainer" containerID="977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371" Dec 05 08:19:05 crc kubenswrapper[4876]: E1205 08:19:05.737372 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371\": container with ID starting with 977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371 not found: ID does not exist" containerID="977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.737425 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371"} err="failed to get container status \"977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371\": rpc error: code = NotFound desc = could not find container \"977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371\": container with ID starting with 977fe8d03fdd060b184f9c789004e9122e92ab891e162acb2c9edf7c73b52371 not found: ID does not exist" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.737462 4876 scope.go:117] "RemoveContainer" containerID="a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7" Dec 05 08:19:05 crc kubenswrapper[4876]: E1205 08:19:05.737896 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7\": container with ID starting with a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7 not found: ID does not exist" containerID="a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.737964 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7"} err="failed to get container status \"a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7\": rpc error: code = NotFound desc = could not find container \"a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7\": container with ID starting with a91fc5921e7be9467f258d106a251a550455d848d50b5580da9ed5264ca659d7 not found: ID does not exist" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.737993 4876 scope.go:117] "RemoveContainer" containerID="cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10" Dec 05 08:19:05 crc kubenswrapper[4876]: E1205 08:19:05.738344 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10\": container with ID starting with cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10 not found: ID does not exist" containerID="cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.738383 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10"} err="failed to get container status \"cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10\": rpc error: code = NotFound desc = could not find container \"cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10\": container with ID starting with cda04a9123488ed1b913c538b8ddfe8cec71e6ccdcbaafa378e9274ee96f2d10 not found: ID does not exist" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.829759 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" path="/var/lib/kubelet/pods/cb460428-9d76-41ca-bebc-6d1edb40d4b1/volumes" Dec 05 08:19:05 crc kubenswrapper[4876]: I1205 08:19:05.830512 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" path="/var/lib/kubelet/pods/e189296a-a3f0-4937-aa28-c9be917fd59c/volumes" Dec 05 08:19:08 crc kubenswrapper[4876]: I1205 08:19:08.219410 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:19:08 crc kubenswrapper[4876]: I1205 08:19:08.219989 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:19:08 crc kubenswrapper[4876]: I1205 08:19:08.220099 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:19:08 crc kubenswrapper[4876]: I1205 08:19:08.221631 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:19:08 crc kubenswrapper[4876]: I1205 08:19:08.221790 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac" gracePeriod=600 Dec 05 08:19:08 crc kubenswrapper[4876]: I1205 08:19:08.617386 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac" exitCode=0 Dec 05 08:19:08 crc kubenswrapper[4876]: I1205 08:19:08.617424 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac"} Dec 05 08:19:08 crc kubenswrapper[4876]: I1205 08:19:08.617720 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"92b82d84a70d1805e62d229730a17c91d2767035530b63f457f1a8e1f0171d73"} Dec 05 08:19:09 crc kubenswrapper[4876]: I1205 08:19:09.766857 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:19:12 crc kubenswrapper[4876]: I1205 08:19:12.025840 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-r8p7s"] Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.059955 4876 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.061291 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e" gracePeriod=15 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.061373 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad" gracePeriod=15 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.061431 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614" gracePeriod=15 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.061345 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96" gracePeriod=15 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.061446 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434" gracePeriod=15 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.063501 4876 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.063850 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="extract-content" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.063877 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="extract-content" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.063959 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerName="extract-content" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.063977 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerName="extract-content" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064001 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064017 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064040 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064057 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064079 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064095 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064113 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerName="extract-utilities" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064129 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerName="extract-utilities" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064149 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69511d2e-da37-424d-bb34-4791b3292c35" containerName="extract-content" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064164 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="69511d2e-da37-424d-bb34-4791b3292c35" containerName="extract-content" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064190 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064205 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064223 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerName="extract-utilities" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064238 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerName="extract-utilities" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064257 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064271 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064297 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="extract-utilities" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064311 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="extract-utilities" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064335 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064349 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064369 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69511d2e-da37-424d-bb34-4791b3292c35" containerName="extract-utilities" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064385 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="69511d2e-da37-424d-bb34-4791b3292c35" containerName="extract-utilities" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064410 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerName="extract-content" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064424 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerName="extract-content" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064441 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064455 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064475 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064489 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064513 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69511d2e-da37-424d-bb34-4791b3292c35" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064527 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="69511d2e-da37-424d-bb34-4791b3292c35" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.064551 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064567 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064838 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064867 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064887 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="69511d2e-da37-424d-bb34-4791b3292c35" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064940 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0e342ab-27c1-442c-926e-f2e9ed1153c5" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064967 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.064992 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.065015 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.065032 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e189296a-a3f0-4937-aa28-c9be917fd59c" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.065055 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.065075 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb460428-9d76-41ca-bebc-6d1edb40d4b1" containerName="registry-server" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.065319 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.065342 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.069809 4876 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.076704 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.082937 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.147627 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.148181 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.148257 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.148384 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.148430 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.148526 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.148592 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.149120 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.175432 4876 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250600 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250672 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250758 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250817 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250838 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250829 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250878 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250936 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.250957 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.251019 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.251049 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.251048 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.251092 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.251114 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.251129 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.251177 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.477228 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:26 crc kubenswrapper[4876]: E1205 08:19:26.498074 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e43e7b90304db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 08:19:26.496302299 +0000 UTC m=+230.984966961,LastTimestamp:2025-12-05 08:19:26.496302299 +0000 UTC m=+230.984966961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.734279 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.736199 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.737209 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96" exitCode=0 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.737240 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614" exitCode=0 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.737251 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad" exitCode=0 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.737263 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434" exitCode=2 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.737324 4876 scope.go:117] "RemoveContainer" containerID="ba7dadbfd6087a1715ab905792ed266717bfbf42fc4491e547e7c7867f893c67" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.739224 4876 generic.go:334] "Generic (PLEG): container finished" podID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" containerID="bff97825b0319fe15bfd495f36ceb6576c3f3476756c6a6752109aaf85816a05" exitCode=0 Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.739302 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d67527e2-6101-4cad-9a5e-3a5a0feed0b9","Type":"ContainerDied","Data":"bff97825b0319fe15bfd495f36ceb6576c3f3476756c6a6752109aaf85816a05"} Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.739988 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:26 crc kubenswrapper[4876]: I1205 08:19:26.741227 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"988a50d6ff6fdd5e2493350529d3f1e99fecf3cd817cfcc8ead59bb000a9ba22"} Dec 05 08:19:27 crc kubenswrapper[4876]: E1205 08:19:27.028464 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e43e7b90304db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 08:19:26.496302299 +0000 UTC m=+230.984966961,LastTimestamp:2025-12-05 08:19:26.496302299 +0000 UTC m=+230.984966961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 08:19:27 crc kubenswrapper[4876]: I1205 08:19:27.750970 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 08:19:27 crc kubenswrapper[4876]: I1205 08:19:27.753809 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745"} Dec 05 08:19:27 crc kubenswrapper[4876]: I1205 08:19:27.755037 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:27 crc kubenswrapper[4876]: E1205 08:19:27.755064 4876 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.046357 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.047182 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.077917 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kubelet-dir\") pod \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.078206 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-var-lock\") pod \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.078357 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kube-api-access\") pod \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\" (UID: \"d67527e2-6101-4cad-9a5e-3a5a0feed0b9\") " Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.078065 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d67527e2-6101-4cad-9a5e-3a5a0feed0b9" (UID: "d67527e2-6101-4cad-9a5e-3a5a0feed0b9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.078280 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-var-lock" (OuterVolumeSpecName: "var-lock") pod "d67527e2-6101-4cad-9a5e-3a5a0feed0b9" (UID: "d67527e2-6101-4cad-9a5e-3a5a0feed0b9"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.078858 4876 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.078983 4876 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.084389 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d67527e2-6101-4cad-9a5e-3a5a0feed0b9" (UID: "d67527e2-6101-4cad-9a5e-3a5a0feed0b9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.179924 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d67527e2-6101-4cad-9a5e-3a5a0feed0b9-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.494534 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.495718 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.496449 4876 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.496872 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.685339 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.685799 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.685924 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.685526 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.685967 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.685995 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.686557 4876 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.686651 4876 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.686728 4876 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.761803 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"d67527e2-6101-4cad-9a5e-3a5a0feed0b9","Type":"ContainerDied","Data":"03139afe014bf515fd6c3137c8a5268f25b0b0ced1a68f2d86705e8f44f2273c"} Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.761854 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03139afe014bf515fd6c3137c8a5268f25b0b0ced1a68f2d86705e8f44f2273c" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.763529 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.765786 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.767165 4876 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e" exitCode=0 Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.767245 4876 scope.go:117] "RemoveContainer" containerID="df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.767297 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.768522 4876 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.792639 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.793255 4876 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.806356 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.807123 4876 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.810559 4876 scope.go:117] "RemoveContainer" containerID="0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.829450 4876 scope.go:117] "RemoveContainer" containerID="5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.837032 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.837703 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.838412 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.838933 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.839488 4876 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.839542 4876 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.839917 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="200ms" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.848751 4876 scope.go:117] "RemoveContainer" containerID="4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.868181 4876 scope.go:117] "RemoveContainer" containerID="9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.885971 4876 scope.go:117] "RemoveContainer" containerID="14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.915505 4876 scope.go:117] "RemoveContainer" containerID="df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.916433 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\": container with ID starting with df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96 not found: ID does not exist" containerID="df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.916463 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96"} err="failed to get container status \"df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\": rpc error: code = NotFound desc = could not find container \"df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96\": container with ID starting with df715e1948f3cf082a3fee664bb6f9d389b0b3c44dc8e68cb7f1c040b04f4e96 not found: ID does not exist" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.916484 4876 scope.go:117] "RemoveContainer" containerID="0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.916934 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\": container with ID starting with 0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614 not found: ID does not exist" containerID="0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.916983 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614"} err="failed to get container status \"0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\": rpc error: code = NotFound desc = could not find container \"0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614\": container with ID starting with 0ce5e2aa885d7da36b891bc86edcc710652f9f4a33987bb41b74c42711acd614 not found: ID does not exist" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.917014 4876 scope.go:117] "RemoveContainer" containerID="5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.917288 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\": container with ID starting with 5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad not found: ID does not exist" containerID="5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.917308 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad"} err="failed to get container status \"5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\": rpc error: code = NotFound desc = could not find container \"5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad\": container with ID starting with 5f21242a60e513688f9f248c1821270d85fda39493b9a832102caba54bc033ad not found: ID does not exist" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.917320 4876 scope.go:117] "RemoveContainer" containerID="4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.917639 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\": container with ID starting with 4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434 not found: ID does not exist" containerID="4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.917658 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434"} err="failed to get container status \"4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\": rpc error: code = NotFound desc = could not find container \"4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434\": container with ID starting with 4d852bd9de615ad7899c5258f69f16c63205e1c366a6091a8d2c34f8e96ba434 not found: ID does not exist" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.917671 4876 scope.go:117] "RemoveContainer" containerID="9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.918104 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\": container with ID starting with 9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e not found: ID does not exist" containerID="9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.918136 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e"} err="failed to get container status \"9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\": rpc error: code = NotFound desc = could not find container \"9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e\": container with ID starting with 9ac1c5fe5515c9907d148e259f69d7fad22edfb3bc6db5c0eab43027718da78e not found: ID does not exist" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.918161 4876 scope.go:117] "RemoveContainer" containerID="14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab" Dec 05 08:19:28 crc kubenswrapper[4876]: E1205 08:19:28.918406 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\": container with ID starting with 14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab not found: ID does not exist" containerID="14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab" Dec 05 08:19:28 crc kubenswrapper[4876]: I1205 08:19:28.918425 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab"} err="failed to get container status \"14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\": rpc error: code = NotFound desc = could not find container \"14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab\": container with ID starting with 14ddbd73eddae18f72e8197409ff69ff26d4ff1a19f93357a3c5680729ddccab not found: ID does not exist" Dec 05 08:19:29 crc kubenswrapper[4876]: E1205 08:19:29.041008 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="400ms" Dec 05 08:19:29 crc kubenswrapper[4876]: E1205 08:19:29.442440 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="800ms" Dec 05 08:19:29 crc kubenswrapper[4876]: I1205 08:19:29.835352 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 05 08:19:30 crc kubenswrapper[4876]: E1205 08:19:30.244563 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="1.6s" Dec 05 08:19:31 crc kubenswrapper[4876]: E1205 08:19:31.845980 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="3.2s" Dec 05 08:19:33 crc kubenswrapper[4876]: E1205 08:19:33.851828 4876 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" volumeName="registry-storage" Dec 05 08:19:35 crc kubenswrapper[4876]: E1205 08:19:35.046692 4876 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="6.4s" Dec 05 08:19:35 crc kubenswrapper[4876]: I1205 08:19:35.829042 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: E1205 08:19:37.030528 4876 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e43e7b90304db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 08:19:26.496302299 +0000 UTC m=+230.984966961,LastTimestamp:2025-12-05 08:19:26.496302299 +0000 UTC m=+230.984966961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.070173 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" containerName="oauth-openshift" containerID="cri-o://afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb" gracePeriod=15 Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.585129 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.586294 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.586763 4876 status_manager.go:851] "Failed to get status for pod" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-r8p7s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750368 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-provider-selection\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750479 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-serving-cert\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750529 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-policies\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750573 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-ocp-branding-template\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750612 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-error\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750649 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-idp-0-file-data\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750685 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-session\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750721 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-router-certs\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750765 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-cliconfig\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750803 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2qtd\" (UniqueName: \"kubernetes.io/projected/4c5a2919-0f9c-4f28-b6e4-ab288f991483-kube-api-access-p2qtd\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750861 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-login\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750922 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-service-ca\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.750967 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-trusted-ca-bundle\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.751001 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-dir\") pod \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\" (UID: \"4c5a2919-0f9c-4f28-b6e4-ab288f991483\") " Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.751286 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.751840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.752478 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.752886 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.753166 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.758668 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.758922 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c5a2919-0f9c-4f28-b6e4-ab288f991483-kube-api-access-p2qtd" (OuterVolumeSpecName: "kube-api-access-p2qtd") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "kube-api-access-p2qtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.759040 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.759650 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.759938 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.760712 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.760950 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.761216 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.761607 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "4c5a2919-0f9c-4f28-b6e4-ab288f991483" (UID: "4c5a2919-0f9c-4f28-b6e4-ab288f991483"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.823616 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.824413 4876 status_manager.go:851] "Failed to get status for pod" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-r8p7s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.825397 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.831790 4876 generic.go:334] "Generic (PLEG): container finished" podID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" containerID="afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb" exitCode=0 Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.831935 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.832790 4876 status_manager.go:851] "Failed to get status for pod" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-r8p7s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.833388 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.839071 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" event={"ID":"4c5a2919-0f9c-4f28-b6e4-ab288f991483","Type":"ContainerDied","Data":"afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb"} Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.839141 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" event={"ID":"4c5a2919-0f9c-4f28-b6e4-ab288f991483","Type":"ContainerDied","Data":"462adc33e726ca1803a13411c15090d226d49bb3f43eceb7128a59801a3b54dd"} Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.839177 4876 scope.go:117] "RemoveContainer" containerID="afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.851936 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.851973 4876 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.851989 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852002 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852017 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852030 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852041 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852053 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852065 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2qtd\" (UniqueName: \"kubernetes.io/projected/4c5a2919-0f9c-4f28-b6e4-ab288f991483-kube-api-access-p2qtd\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852076 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852089 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852147 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852162 4876 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4c5a2919-0f9c-4f28-b6e4-ab288f991483-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.852203 4876 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4c5a2919-0f9c-4f28-b6e4-ab288f991483-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.869993 4876 scope.go:117] "RemoveContainer" containerID="afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb" Dec 05 08:19:37 crc kubenswrapper[4876]: E1205 08:19:37.870836 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb\": container with ID starting with afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb not found: ID does not exist" containerID="afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.871125 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb"} err="failed to get container status \"afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb\": rpc error: code = NotFound desc = could not find container \"afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb\": container with ID starting with afcc0fd3019276e62da15940ec30e5b03708bf95809c694119568a3d4e1e4ddb not found: ID does not exist" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.879827 4876 status_manager.go:851] "Failed to get status for pod" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-r8p7s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.880608 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.885064 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.885093 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:37 crc kubenswrapper[4876]: E1205 08:19:37.885333 4876 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:37 crc kubenswrapper[4876]: I1205 08:19:37.885803 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:38 crc kubenswrapper[4876]: I1205 08:19:38.848530 4876 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="30ec01bf00577886aef43fa82ca7c8d712db87919027d36994c0eba7a5d14306" exitCode=0 Dec 05 08:19:38 crc kubenswrapper[4876]: I1205 08:19:38.849403 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"30ec01bf00577886aef43fa82ca7c8d712db87919027d36994c0eba7a5d14306"} Dec 05 08:19:38 crc kubenswrapper[4876]: I1205 08:19:38.849452 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1c1de57a51f79042f403d2aca1ffe5469eb64faf9f94f31a08113b6dd5faeefb"} Dec 05 08:19:38 crc kubenswrapper[4876]: I1205 08:19:38.850061 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:38 crc kubenswrapper[4876]: I1205 08:19:38.850089 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:38 crc kubenswrapper[4876]: E1205 08:19:38.850727 4876 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:38 crc kubenswrapper[4876]: I1205 08:19:38.850860 4876 status_manager.go:851] "Failed to get status for pod" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" pod="openshift-authentication/oauth-openshift-558db77b4-r8p7s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-r8p7s\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:38 crc kubenswrapper[4876]: I1205 08:19:38.851458 4876 status_manager.go:851] "Failed to get status for pod" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.234:6443: connect: connection refused" Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.264709 4876 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.264813 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.864772 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.865164 4876 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739" exitCode=1 Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.865245 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739"} Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.865746 4876 scope.go:117] "RemoveContainer" containerID="0f1afc8caa8d82fa6a174b4f8a86db1588b557a8a0a3710eeea07fa8ce7c6739" Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.869161 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"46c9d9b60c4d1b7848c46f403d773340615bdaa6cc650461b09ea7170bf7a6f6"} Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.869184 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ebc35f3d711ad87d74fa30f0c7909c72347b1f125af033f8182f1c14170e7352"} Dec 05 08:19:39 crc kubenswrapper[4876]: I1205 08:19:39.869194 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"877af5feb18527ee384912e922a8eb5a7f8bbe9abfec1b633a898dcd9c613886"} Dec 05 08:19:40 crc kubenswrapper[4876]: I1205 08:19:40.876622 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"31a4802b9bfc2543e6d2ec3b51ee3c501b360a154492cafba11b2f0f2bf6d584"} Dec 05 08:19:40 crc kubenswrapper[4876]: I1205 08:19:40.876684 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cfee19a3ff5014ca76a5c63c8e2076ffa4f4808be9921d43ce82ea70dd4ae230"} Dec 05 08:19:40 crc kubenswrapper[4876]: I1205 08:19:40.876874 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:40 crc kubenswrapper[4876]: I1205 08:19:40.877138 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:40 crc kubenswrapper[4876]: I1205 08:19:40.877181 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:40 crc kubenswrapper[4876]: I1205 08:19:40.880473 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 08:19:40 crc kubenswrapper[4876]: I1205 08:19:40.880521 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ed92d45fd0de3eb879756226cead7f6810770b1841a0db4cecaead14d0aac505"} Dec 05 08:19:42 crc kubenswrapper[4876]: I1205 08:19:42.887199 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:42 crc kubenswrapper[4876]: I1205 08:19:42.887486 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:42 crc kubenswrapper[4876]: I1205 08:19:42.892868 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:43 crc kubenswrapper[4876]: I1205 08:19:43.224041 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:19:43 crc kubenswrapper[4876]: I1205 08:19:43.228051 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:19:43 crc kubenswrapper[4876]: I1205 08:19:43.897803 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:19:45 crc kubenswrapper[4876]: I1205 08:19:45.887447 4876 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:45 crc kubenswrapper[4876]: I1205 08:19:45.908147 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:45 crc kubenswrapper[4876]: I1205 08:19:45.908236 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:45 crc kubenswrapper[4876]: I1205 08:19:45.915459 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:19:45 crc kubenswrapper[4876]: I1205 08:19:45.919575 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="fbee5dca-290c-47e8-a9a8-ece7f6166cc0" Dec 05 08:19:46 crc kubenswrapper[4876]: I1205 08:19:46.912072 4876 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:46 crc kubenswrapper[4876]: I1205 08:19:46.912102 4876 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4846bdfd-e3f9-43b1-bc83-b8d155ce793d" Dec 05 08:19:55 crc kubenswrapper[4876]: I1205 08:19:55.843461 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="fbee5dca-290c-47e8-a9a8-ece7f6166cc0" Dec 05 08:19:57 crc kubenswrapper[4876]: I1205 08:19:57.171243 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 08:19:57 crc kubenswrapper[4876]: I1205 08:19:57.538532 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 08:19:57 crc kubenswrapper[4876]: I1205 08:19:57.990791 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 08:19:58 crc kubenswrapper[4876]: I1205 08:19:58.064183 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 08:19:58 crc kubenswrapper[4876]: I1205 08:19:58.139996 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 08:19:58 crc kubenswrapper[4876]: I1205 08:19:58.437436 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 08:19:58 crc kubenswrapper[4876]: I1205 08:19:58.633593 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 08:19:58 crc kubenswrapper[4876]: I1205 08:19:58.667321 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 08:19:58 crc kubenswrapper[4876]: I1205 08:19:58.688185 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 08:19:58 crc kubenswrapper[4876]: I1205 08:19:58.688762 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 08:19:58 crc kubenswrapper[4876]: I1205 08:19:58.784858 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.269037 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.274032 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.274451 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.364012 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.430374 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.536410 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.541056 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.622789 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.663880 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.732075 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.772517 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.813529 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.831500 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.903588 4876 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 08:19:59 crc kubenswrapper[4876]: I1205 08:19:59.907510 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.017872 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.060535 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.163481 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.395855 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.417569 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.471283 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.506596 4876 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.513881 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-r8p7s","openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.514008 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.521767 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.545576 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.545599 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.545574086 podStartE2EDuration="15.545574086s" podCreationTimestamp="2025-12-05 08:19:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:20:00.541390298 +0000 UTC m=+265.030054980" watchObservedRunningTime="2025-12-05 08:20:00.545574086 +0000 UTC m=+265.034238748" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.598012 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.642822 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.675796 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.717098 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.782755 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.839151 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.879051 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.905075 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 08:20:00 crc kubenswrapper[4876]: I1205 08:20:00.981739 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.056487 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.166026 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.296218 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.359528 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.467523 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.495675 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.605883 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.690886 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.805659 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.808743 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.837777 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" path="/var/lib/kubelet/pods/4c5a2919-0f9c-4f28-b6e4-ab288f991483/volumes" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.839214 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.848201 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.970325 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 08:20:01 crc kubenswrapper[4876]: I1205 08:20:01.979362 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.134577 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.196545 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.217109 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.242652 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.276074 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.298878 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.305350 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.532324 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.533043 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.543703 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.644265 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-69b74fc85f-qv5qf"] Dec 05 08:20:02 crc kubenswrapper[4876]: E1205 08:20:02.644491 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" containerName="oauth-openshift" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.644505 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" containerName="oauth-openshift" Dec 05 08:20:02 crc kubenswrapper[4876]: E1205 08:20:02.644529 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" containerName="installer" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.644537 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" containerName="installer" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.644641 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5a2919-0f9c-4f28-b6e4-ab288f991483" containerName="oauth-openshift" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.644660 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67527e2-6101-4cad-9a5e-3a5a0feed0b9" containerName="installer" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.645215 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.654497 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.654509 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.654823 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.654472 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.655497 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.656179 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.656419 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.656821 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.657482 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.657533 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.657630 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.659624 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.660841 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-69b74fc85f-qv5qf"] Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.668429 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.677239 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.687727 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.712508 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.768023 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.843810 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.843876 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.843952 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfjnx\" (UniqueName: \"kubernetes.io/projected/db3d454f-a614-4c71-901a-e1bcb613ecf6-kube-api-access-rfjnx\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844054 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-error\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844096 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/db3d454f-a614-4c71-901a-e1bcb613ecf6-audit-dir\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844126 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-audit-policies\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844153 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844182 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-service-ca\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844213 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844251 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844293 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844324 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-login\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844423 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-router-certs\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.844466 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-session\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.945682 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.946699 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.946731 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfjnx\" (UniqueName: \"kubernetes.io/projected/db3d454f-a614-4c71-901a-e1bcb613ecf6-kube-api-access-rfjnx\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.946787 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-error\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.946811 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/db3d454f-a614-4c71-901a-e1bcb613ecf6-audit-dir\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.946886 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/db3d454f-a614-4c71-901a-e1bcb613ecf6-audit-dir\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.946940 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-audit-policies\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.946963 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.947748 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.947783 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-audit-policies\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.947848 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-service-ca\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.947878 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.947920 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.947932 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-service-ca\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.947954 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.947981 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-login\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.948012 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-router-certs\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.948035 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-session\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.948484 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.948618 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.951891 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-login\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.953385 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.958188 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.959858 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.968813 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-template-error\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.969869 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.970944 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.971746 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.972403 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfjnx\" (UniqueName: \"kubernetes.io/projected/db3d454f-a614-4c71-901a-e1bcb613ecf6-kube-api-access-rfjnx\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.974352 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-session\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.975972 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/db3d454f-a614-4c71-901a-e1bcb613ecf6-v4-0-config-system-router-certs\") pod \"oauth-openshift-69b74fc85f-qv5qf\" (UID: \"db3d454f-a614-4c71-901a-e1bcb613ecf6\") " pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:02 crc kubenswrapper[4876]: I1205 08:20:02.984160 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.011139 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.105703 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.202533 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.260426 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.288638 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.318454 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.321731 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.369493 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.380737 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.385589 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.528503 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.549433 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.650758 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.709016 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.789690 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.824094 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.855775 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.955288 4876 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 08:20:03 crc kubenswrapper[4876]: I1205 08:20:03.961075 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.088600 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.209846 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.210341 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.214446 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.222641 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.251794 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.463247 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.474523 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.499245 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.733150 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.745199 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.763062 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.772126 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 08:20:04 crc kubenswrapper[4876]: I1205 08:20:04.934307 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.082388 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.108516 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.159895 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.185096 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.185121 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.205319 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.385974 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.446621 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.473655 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.505406 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.564360 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.569627 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.745959 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.789089 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-69b74fc85f-qv5qf"] Dec 05 08:20:05 crc kubenswrapper[4876]: W1205 08:20:05.798278 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb3d454f_a614_4c71_901a_e1bcb613ecf6.slice/crio-fbfea7c02a40a918900847a75916f96ab175e34b64234c6c704dec9b060ae322 WatchSource:0}: Error finding container fbfea7c02a40a918900847a75916f96ab175e34b64234c6c704dec9b060ae322: Status 404 returned error can't find the container with id fbfea7c02a40a918900847a75916f96ab175e34b64234c6c704dec9b060ae322 Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.883045 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.909519 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 08:20:05 crc kubenswrapper[4876]: I1205 08:20:05.926605 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.064075 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" event={"ID":"db3d454f-a614-4c71-901a-e1bcb613ecf6","Type":"ContainerStarted","Data":"fbfea7c02a40a918900847a75916f96ab175e34b64234c6c704dec9b060ae322"} Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.103086 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.168285 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.193483 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.220474 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.233863 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.292361 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.295147 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.297940 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.302943 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.344845 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.404339 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.409021 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.451514 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.453890 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.524718 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.542066 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.554589 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.557571 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.608378 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.659064 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.664742 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.681479 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.755278 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.842072 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.880719 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.935842 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 08:20:06 crc kubenswrapper[4876]: I1205 08:20:06.940446 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.062018 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.071210 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.073639 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" event={"ID":"db3d454f-a614-4c71-901a-e1bcb613ecf6","Type":"ContainerStarted","Data":"16d0e11a46d8b3c512d48811c9e5182bdc103300adadd7bbf98ab7fff9f706ca"} Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.074239 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.081341 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.105475 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-69b74fc85f-qv5qf" podStartSLOduration=55.105452286 podStartE2EDuration="55.105452286s" podCreationTimestamp="2025-12-05 08:19:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:20:07.100120666 +0000 UTC m=+271.588785348" watchObservedRunningTime="2025-12-05 08:20:07.105452286 +0000 UTC m=+271.594116938" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.121643 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.149720 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.215465 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.280360 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.409185 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.479973 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.483591 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.594373 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.613058 4876 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.654306 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.704454 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.722989 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.723001 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.815340 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.881228 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.882101 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.912943 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 08:20:07 crc kubenswrapper[4876]: I1205 08:20:07.962416 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.145435 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.165888 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.219805 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.236577 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.246479 4876 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.246878 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745" gracePeriod=5 Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.370572 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.399293 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.421194 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.428652 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.457058 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.504044 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.507925 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.524037 4876 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.631708 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.822107 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.862711 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.877130 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.914520 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.915492 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.941047 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 08:20:08 crc kubenswrapper[4876]: I1205 08:20:08.944192 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 08:20:09 crc kubenswrapper[4876]: I1205 08:20:09.143747 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 08:20:09 crc kubenswrapper[4876]: I1205 08:20:09.153248 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 08:20:09 crc kubenswrapper[4876]: I1205 08:20:09.211207 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 08:20:09 crc kubenswrapper[4876]: I1205 08:20:09.219000 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 08:20:09 crc kubenswrapper[4876]: I1205 08:20:09.385835 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 08:20:09 crc kubenswrapper[4876]: I1205 08:20:09.570044 4876 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 08:20:09 crc kubenswrapper[4876]: I1205 08:20:09.577977 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.023345 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.116107 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.144981 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.159700 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.247983 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.253540 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.265828 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.312122 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.340874 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.456356 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.525722 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.752527 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.758703 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.823460 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.824978 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.878669 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.930732 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 08:20:10 crc kubenswrapper[4876]: I1205 08:20:10.960736 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.051749 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.107573 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.139409 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.174176 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.242231 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.248475 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.340971 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.414485 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.532531 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.904974 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 08:20:11 crc kubenswrapper[4876]: I1205 08:20:11.990411 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 08:20:12 crc kubenswrapper[4876]: I1205 08:20:12.047052 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 08:20:12 crc kubenswrapper[4876]: I1205 08:20:12.047776 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 08:20:12 crc kubenswrapper[4876]: I1205 08:20:12.105083 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 08:20:12 crc kubenswrapper[4876]: I1205 08:20:12.270084 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 08:20:12 crc kubenswrapper[4876]: I1205 08:20:12.388068 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 08:20:12 crc kubenswrapper[4876]: I1205 08:20:12.439343 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 08:20:12 crc kubenswrapper[4876]: I1205 08:20:12.491490 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 08:20:12 crc kubenswrapper[4876]: I1205 08:20:12.990722 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.032091 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.347731 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.759442 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.859773 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.859837 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908141 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908205 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908240 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908288 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908313 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908376 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908448 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908470 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908529 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.908816 4876 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.909040 4876 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.909070 4876 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:13 crc kubenswrapper[4876]: I1205 08:20:13.916472 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.009972 4876 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.010032 4876 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.110741 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.111056 4876 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745" exitCode=137 Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.111161 4876 scope.go:117] "RemoveContainer" containerID="c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.111176 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.135148 4876 scope.go:117] "RemoveContainer" containerID="c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745" Dec 05 08:20:14 crc kubenswrapper[4876]: E1205 08:20:14.135912 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745\": container with ID starting with c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745 not found: ID does not exist" containerID="c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.135980 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745"} err="failed to get container status \"c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745\": rpc error: code = NotFound desc = could not find container \"c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745\": container with ID starting with c1e0b741e815a5e924fd32acbc1e0472d6853825b23f397ca37ae0824642b745 not found: ID does not exist" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.555388 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 08:20:14 crc kubenswrapper[4876]: I1205 08:20:14.681628 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 08:20:15 crc kubenswrapper[4876]: I1205 08:20:15.673385 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 08:20:15 crc kubenswrapper[4876]: I1205 08:20:15.836841 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 05 08:20:28 crc kubenswrapper[4876]: I1205 08:20:28.195229 4876 generic.go:334] "Generic (PLEG): container finished" podID="36087227-ee02-4dd0-b801-2ced90e4173f" containerID="bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e" exitCode=0 Dec 05 08:20:28 crc kubenswrapper[4876]: I1205 08:20:28.195313 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" event={"ID":"36087227-ee02-4dd0-b801-2ced90e4173f","Type":"ContainerDied","Data":"bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e"} Dec 05 08:20:28 crc kubenswrapper[4876]: I1205 08:20:28.196524 4876 scope.go:117] "RemoveContainer" containerID="bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e" Dec 05 08:20:29 crc kubenswrapper[4876]: I1205 08:20:29.211221 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" event={"ID":"36087227-ee02-4dd0-b801-2ced90e4173f","Type":"ContainerStarted","Data":"bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600"} Dec 05 08:20:29 crc kubenswrapper[4876]: I1205 08:20:29.211795 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:20:29 crc kubenswrapper[4876]: I1205 08:20:29.213615 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.509711 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6nzhg"] Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.510608 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" podUID="b15da8e1-2f29-41fa-8dae-dce111e40262" containerName="controller-manager" containerID="cri-o://5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a" gracePeriod=30 Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.648198 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z"] Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.648392 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" podUID="859647d8-54ea-4892-ae54-2179375e1ae0" containerName="route-controller-manager" containerID="cri-o://0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988" gracePeriod=30 Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.911917 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.985276 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.993055 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bqpm\" (UniqueName: \"kubernetes.io/projected/b15da8e1-2f29-41fa-8dae-dce111e40262-kube-api-access-4bqpm\") pod \"b15da8e1-2f29-41fa-8dae-dce111e40262\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.993108 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-proxy-ca-bundles\") pod \"b15da8e1-2f29-41fa-8dae-dce111e40262\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.993129 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-config\") pod \"b15da8e1-2f29-41fa-8dae-dce111e40262\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.993175 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15da8e1-2f29-41fa-8dae-dce111e40262-serving-cert\") pod \"b15da8e1-2f29-41fa-8dae-dce111e40262\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.993192 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-client-ca\") pod \"b15da8e1-2f29-41fa-8dae-dce111e40262\" (UID: \"b15da8e1-2f29-41fa-8dae-dce111e40262\") " Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.994717 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-config" (OuterVolumeSpecName: "config") pod "b15da8e1-2f29-41fa-8dae-dce111e40262" (UID: "b15da8e1-2f29-41fa-8dae-dce111e40262"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.994889 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b15da8e1-2f29-41fa-8dae-dce111e40262" (UID: "b15da8e1-2f29-41fa-8dae-dce111e40262"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.995063 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-client-ca" (OuterVolumeSpecName: "client-ca") pod "b15da8e1-2f29-41fa-8dae-dce111e40262" (UID: "b15da8e1-2f29-41fa-8dae-dce111e40262"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:20:34 crc kubenswrapper[4876]: I1205 08:20:34.999059 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b15da8e1-2f29-41fa-8dae-dce111e40262-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b15da8e1-2f29-41fa-8dae-dce111e40262" (UID: "b15da8e1-2f29-41fa-8dae-dce111e40262"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.047525 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b15da8e1-2f29-41fa-8dae-dce111e40262-kube-api-access-4bqpm" (OuterVolumeSpecName: "kube-api-access-4bqpm") pod "b15da8e1-2f29-41fa-8dae-dce111e40262" (UID: "b15da8e1-2f29-41fa-8dae-dce111e40262"). InnerVolumeSpecName "kube-api-access-4bqpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.095359 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-client-ca\") pod \"859647d8-54ea-4892-ae54-2179375e1ae0\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.095475 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-config\") pod \"859647d8-54ea-4892-ae54-2179375e1ae0\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.095511 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/859647d8-54ea-4892-ae54-2179375e1ae0-serving-cert\") pod \"859647d8-54ea-4892-ae54-2179375e1ae0\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.095571 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9qng\" (UniqueName: \"kubernetes.io/projected/859647d8-54ea-4892-ae54-2179375e1ae0-kube-api-access-x9qng\") pod \"859647d8-54ea-4892-ae54-2179375e1ae0\" (UID: \"859647d8-54ea-4892-ae54-2179375e1ae0\") " Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.095932 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15da8e1-2f29-41fa-8dae-dce111e40262-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.095961 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.095980 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bqpm\" (UniqueName: \"kubernetes.io/projected/b15da8e1-2f29-41fa-8dae-dce111e40262-kube-api-access-4bqpm\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.095998 4876 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.096015 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15da8e1-2f29-41fa-8dae-dce111e40262-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.096840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-client-ca" (OuterVolumeSpecName: "client-ca") pod "859647d8-54ea-4892-ae54-2179375e1ae0" (UID: "859647d8-54ea-4892-ae54-2179375e1ae0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.096867 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-config" (OuterVolumeSpecName: "config") pod "859647d8-54ea-4892-ae54-2179375e1ae0" (UID: "859647d8-54ea-4892-ae54-2179375e1ae0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.099986 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/859647d8-54ea-4892-ae54-2179375e1ae0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "859647d8-54ea-4892-ae54-2179375e1ae0" (UID: "859647d8-54ea-4892-ae54-2179375e1ae0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.100529 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/859647d8-54ea-4892-ae54-2179375e1ae0-kube-api-access-x9qng" (OuterVolumeSpecName: "kube-api-access-x9qng") pod "859647d8-54ea-4892-ae54-2179375e1ae0" (UID: "859647d8-54ea-4892-ae54-2179375e1ae0"). InnerVolumeSpecName "kube-api-access-x9qng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.196672 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.196716 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/859647d8-54ea-4892-ae54-2179375e1ae0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.196743 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9qng\" (UniqueName: \"kubernetes.io/projected/859647d8-54ea-4892-ae54-2179375e1ae0-kube-api-access-x9qng\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.196761 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/859647d8-54ea-4892-ae54-2179375e1ae0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.251662 4876 generic.go:334] "Generic (PLEG): container finished" podID="859647d8-54ea-4892-ae54-2179375e1ae0" containerID="0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988" exitCode=0 Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.251717 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.251729 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" event={"ID":"859647d8-54ea-4892-ae54-2179375e1ae0","Type":"ContainerDied","Data":"0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988"} Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.251788 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z" event={"ID":"859647d8-54ea-4892-ae54-2179375e1ae0","Type":"ContainerDied","Data":"8112206fa85e10508f0990c604e387f6c984eb21f9b7d8e9442fdbdaa7ea7090"} Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.251830 4876 scope.go:117] "RemoveContainer" containerID="0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.254846 4876 generic.go:334] "Generic (PLEG): container finished" podID="b15da8e1-2f29-41fa-8dae-dce111e40262" containerID="5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a" exitCode=0 Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.254890 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" event={"ID":"b15da8e1-2f29-41fa-8dae-dce111e40262","Type":"ContainerDied","Data":"5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a"} Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.254952 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" event={"ID":"b15da8e1-2f29-41fa-8dae-dce111e40262","Type":"ContainerDied","Data":"7b55d99eb74589b2033f67cd7863aec6c07c315cdf1ccb3cca565a50c41abf1c"} Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.255014 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6nzhg" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.278801 4876 scope.go:117] "RemoveContainer" containerID="0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988" Dec 05 08:20:35 crc kubenswrapper[4876]: E1205 08:20:35.279306 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988\": container with ID starting with 0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988 not found: ID does not exist" containerID="0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.279356 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988"} err="failed to get container status \"0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988\": rpc error: code = NotFound desc = could not find container \"0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988\": container with ID starting with 0c4d2366439cc688a1adb1180895502733887caff22beb170fff1ae0a5a29988 not found: ID does not exist" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.279387 4876 scope.go:117] "RemoveContainer" containerID="5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.293811 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6nzhg"] Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.300021 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6nzhg"] Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.306143 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z"] Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.309650 4876 scope.go:117] "RemoveContainer" containerID="5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a" Dec 05 08:20:35 crc kubenswrapper[4876]: E1205 08:20:35.310199 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a\": container with ID starting with 5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a not found: ID does not exist" containerID="5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.310396 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a"} err="failed to get container status \"5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a\": rpc error: code = NotFound desc = could not find container \"5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a\": container with ID starting with 5b93f4fa0cf087eb34a3077408a656c221418a0c87a81fe8017b0d14fa2be56a not found: ID does not exist" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.310602 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-2c65z"] Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.836945 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="859647d8-54ea-4892-ae54-2179375e1ae0" path="/var/lib/kubelet/pods/859647d8-54ea-4892-ae54-2179375e1ae0/volumes" Dec 05 08:20:35 crc kubenswrapper[4876]: I1205 08:20:35.838163 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b15da8e1-2f29-41fa-8dae-dce111e40262" path="/var/lib/kubelet/pods/b15da8e1-2f29-41fa-8dae-dce111e40262/volumes" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.472712 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb"] Dec 05 08:20:36 crc kubenswrapper[4876]: E1205 08:20:36.473722 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="859647d8-54ea-4892-ae54-2179375e1ae0" containerName="route-controller-manager" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.473828 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="859647d8-54ea-4892-ae54-2179375e1ae0" containerName="route-controller-manager" Dec 05 08:20:36 crc kubenswrapper[4876]: E1205 08:20:36.473932 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15da8e1-2f29-41fa-8dae-dce111e40262" containerName="controller-manager" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.474011 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15da8e1-2f29-41fa-8dae-dce111e40262" containerName="controller-manager" Dec 05 08:20:36 crc kubenswrapper[4876]: E1205 08:20:36.474104 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.474187 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.474393 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="859647d8-54ea-4892-ae54-2179375e1ae0" containerName="route-controller-manager" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.474479 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.474555 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b15da8e1-2f29-41fa-8dae-dce111e40262" containerName="controller-manager" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.475114 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.476202 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm"] Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.477034 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.483300 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.483675 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.483955 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.484041 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.484164 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.484332 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.485085 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.486253 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.486505 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.486746 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.487042 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.487452 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.489652 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm"] Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.502241 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.507446 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb"] Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517135 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-proxy-ca-bundles\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517192 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-client-ca\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517241 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-client-ca\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517276 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-config\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517300 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82kxf\" (UniqueName: \"kubernetes.io/projected/933563d1-e807-464b-9c08-eca45add16c2-kube-api-access-82kxf\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517354 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a785619-6b17-4496-bdaa-95244ef2bfb5-serving-cert\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517375 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgz9t\" (UniqueName: \"kubernetes.io/projected/5a785619-6b17-4496-bdaa-95244ef2bfb5-kube-api-access-vgz9t\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517396 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-config\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.517438 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/933563d1-e807-464b-9c08-eca45add16c2-serving-cert\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.618975 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a785619-6b17-4496-bdaa-95244ef2bfb5-serving-cert\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.619022 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgz9t\" (UniqueName: \"kubernetes.io/projected/5a785619-6b17-4496-bdaa-95244ef2bfb5-kube-api-access-vgz9t\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.619047 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-config\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.619084 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/933563d1-e807-464b-9c08-eca45add16c2-serving-cert\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.619104 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-proxy-ca-bundles\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.619125 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-client-ca\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.619142 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-client-ca\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.619165 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-config\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.619180 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82kxf\" (UniqueName: \"kubernetes.io/projected/933563d1-e807-464b-9c08-eca45add16c2-kube-api-access-82kxf\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.620680 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-client-ca\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.620796 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-proxy-ca-bundles\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.621886 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-client-ca\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.622200 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/933563d1-e807-464b-9c08-eca45add16c2-config\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.622440 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-config\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.628056 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/933563d1-e807-464b-9c08-eca45add16c2-serving-cert\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.634701 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a785619-6b17-4496-bdaa-95244ef2bfb5-serving-cert\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.639186 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgz9t\" (UniqueName: \"kubernetes.io/projected/5a785619-6b17-4496-bdaa-95244ef2bfb5-kube-api-access-vgz9t\") pod \"route-controller-manager-697f44cb6c-m7vvm\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.649089 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82kxf\" (UniqueName: \"kubernetes.io/projected/933563d1-e807-464b-9c08-eca45add16c2-kube-api-access-82kxf\") pod \"controller-manager-6ddc46bb4f-kmqxb\" (UID: \"933563d1-e807-464b-9c08-eca45add16c2\") " pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.813672 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:36 crc kubenswrapper[4876]: I1205 08:20:36.827093 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.048678 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb"] Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.093498 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm"] Dec 05 08:20:37 crc kubenswrapper[4876]: W1205 08:20:37.104599 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a785619_6b17_4496_bdaa_95244ef2bfb5.slice/crio-d1eab1d036929cbc6aba70579b3a17543a3c1914b75f034d571fd95e9c636852 WatchSource:0}: Error finding container d1eab1d036929cbc6aba70579b3a17543a3c1914b75f034d571fd95e9c636852: Status 404 returned error can't find the container with id d1eab1d036929cbc6aba70579b3a17543a3c1914b75f034d571fd95e9c636852 Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.266934 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" event={"ID":"933563d1-e807-464b-9c08-eca45add16c2","Type":"ContainerStarted","Data":"08d08dabd331aec8b242cb0b6307caaea0f56146e9486b5e1d499dcd25866f9f"} Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.267479 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.267598 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" event={"ID":"933563d1-e807-464b-9c08-eca45add16c2","Type":"ContainerStarted","Data":"429e3aa99c86c733db9be009759fb370f51748f0321d34b4060dc983619e65f7"} Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.268297 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" event={"ID":"5a785619-6b17-4496-bdaa-95244ef2bfb5","Type":"ContainerStarted","Data":"5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c"} Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.268397 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" event={"ID":"5a785619-6b17-4496-bdaa-95244ef2bfb5","Type":"ContainerStarted","Data":"d1eab1d036929cbc6aba70579b3a17543a3c1914b75f034d571fd95e9c636852"} Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.268694 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.269729 4876 patch_prober.go:28] interesting pod/route-controller-manager-697f44cb6c-m7vvm container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.269817 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" podUID="5a785619-6b17-4496-bdaa-95244ef2bfb5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.276106 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.288604 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6ddc46bb4f-kmqxb" podStartSLOduration=3.288576027 podStartE2EDuration="3.288576027s" podCreationTimestamp="2025-12-05 08:20:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:20:37.283636728 +0000 UTC m=+301.772301350" watchObservedRunningTime="2025-12-05 08:20:37.288576027 +0000 UTC m=+301.777240679" Dec 05 08:20:37 crc kubenswrapper[4876]: I1205 08:20:37.310965 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" podStartSLOduration=3.310945297 podStartE2EDuration="3.310945297s" podCreationTimestamp="2025-12-05 08:20:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:20:37.308271372 +0000 UTC m=+301.796935994" watchObservedRunningTime="2025-12-05 08:20:37.310945297 +0000 UTC m=+301.799609959" Dec 05 08:20:38 crc kubenswrapper[4876]: I1205 08:20:38.280982 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.240201 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm"] Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.241260 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" podUID="5a785619-6b17-4496-bdaa-95244ef2bfb5" containerName="route-controller-manager" containerID="cri-o://5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c" gracePeriod=30 Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.648321 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.681605 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a785619-6b17-4496-bdaa-95244ef2bfb5-serving-cert\") pod \"5a785619-6b17-4496-bdaa-95244ef2bfb5\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.681666 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-config\") pod \"5a785619-6b17-4496-bdaa-95244ef2bfb5\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.681734 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgz9t\" (UniqueName: \"kubernetes.io/projected/5a785619-6b17-4496-bdaa-95244ef2bfb5-kube-api-access-vgz9t\") pod \"5a785619-6b17-4496-bdaa-95244ef2bfb5\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.681753 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-client-ca\") pod \"5a785619-6b17-4496-bdaa-95244ef2bfb5\" (UID: \"5a785619-6b17-4496-bdaa-95244ef2bfb5\") " Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.682425 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-client-ca" (OuterVolumeSpecName: "client-ca") pod "5a785619-6b17-4496-bdaa-95244ef2bfb5" (UID: "5a785619-6b17-4496-bdaa-95244ef2bfb5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.682454 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-config" (OuterVolumeSpecName: "config") pod "5a785619-6b17-4496-bdaa-95244ef2bfb5" (UID: "5a785619-6b17-4496-bdaa-95244ef2bfb5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.688023 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a785619-6b17-4496-bdaa-95244ef2bfb5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5a785619-6b17-4496-bdaa-95244ef2bfb5" (UID: "5a785619-6b17-4496-bdaa-95244ef2bfb5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.688644 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a785619-6b17-4496-bdaa-95244ef2bfb5-kube-api-access-vgz9t" (OuterVolumeSpecName: "kube-api-access-vgz9t") pod "5a785619-6b17-4496-bdaa-95244ef2bfb5" (UID: "5a785619-6b17-4496-bdaa-95244ef2bfb5"). InnerVolumeSpecName "kube-api-access-vgz9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.783147 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a785619-6b17-4496-bdaa-95244ef2bfb5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.783201 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.783211 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgz9t\" (UniqueName: \"kubernetes.io/projected/5a785619-6b17-4496-bdaa-95244ef2bfb5-kube-api-access-vgz9t\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:41 crc kubenswrapper[4876]: I1205 08:20:41.783224 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a785619-6b17-4496-bdaa-95244ef2bfb5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.298785 4876 generic.go:334] "Generic (PLEG): container finished" podID="5a785619-6b17-4496-bdaa-95244ef2bfb5" containerID="5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c" exitCode=0 Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.298881 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" event={"ID":"5a785619-6b17-4496-bdaa-95244ef2bfb5","Type":"ContainerDied","Data":"5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c"} Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.298891 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.298974 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm" event={"ID":"5a785619-6b17-4496-bdaa-95244ef2bfb5","Type":"ContainerDied","Data":"d1eab1d036929cbc6aba70579b3a17543a3c1914b75f034d571fd95e9c636852"} Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.298999 4876 scope.go:117] "RemoveContainer" containerID="5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.323972 4876 scope.go:117] "RemoveContainer" containerID="5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.324082 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm"] Dec 05 08:20:42 crc kubenswrapper[4876]: E1205 08:20:42.324339 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c\": container with ID starting with 5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c not found: ID does not exist" containerID="5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.324369 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c"} err="failed to get container status \"5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c\": rpc error: code = NotFound desc = could not find container \"5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c\": container with ID starting with 5a4c73e5cbb9a74252cb76d77197b4d06c8c6d5a1dad29c8d0a76cb3b1f9696c not found: ID does not exist" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.329827 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-m7vvm"] Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.483636 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs"] Dec 05 08:20:42 crc kubenswrapper[4876]: E1205 08:20:42.484547 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a785619-6b17-4496-bdaa-95244ef2bfb5" containerName="route-controller-manager" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.484762 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a785619-6b17-4496-bdaa-95244ef2bfb5" containerName="route-controller-manager" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.485238 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a785619-6b17-4496-bdaa-95244ef2bfb5" containerName="route-controller-manager" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.486252 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.490009 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.490063 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.490426 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.490609 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.490746 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.490768 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.507014 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs"] Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.593296 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-config\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.593361 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgf9m\" (UniqueName: \"kubernetes.io/projected/68559e3a-898b-48f2-8a86-362dcfe74e78-kube-api-access-xgf9m\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.593625 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-client-ca\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.593743 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68559e3a-898b-48f2-8a86-362dcfe74e78-serving-cert\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.695348 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-config\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.697090 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgf9m\" (UniqueName: \"kubernetes.io/projected/68559e3a-898b-48f2-8a86-362dcfe74e78-kube-api-access-xgf9m\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.697141 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-config\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.697174 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-client-ca\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.697682 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68559e3a-898b-48f2-8a86-362dcfe74e78-serving-cert\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.697836 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-client-ca\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.703814 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68559e3a-898b-48f2-8a86-362dcfe74e78-serving-cert\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.727555 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgf9m\" (UniqueName: \"kubernetes.io/projected/68559e3a-898b-48f2-8a86-362dcfe74e78-kube-api-access-xgf9m\") pod \"route-controller-manager-8678c97bff-lhzhs\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:42 crc kubenswrapper[4876]: I1205 08:20:42.858516 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:43 crc kubenswrapper[4876]: I1205 08:20:43.289367 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs"] Dec 05 08:20:43 crc kubenswrapper[4876]: I1205 08:20:43.311748 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" event={"ID":"68559e3a-898b-48f2-8a86-362dcfe74e78","Type":"ContainerStarted","Data":"da8fda97687c9e89bb776eff39c9082ff9e7eef5c2c9af7dbefab6d24c4bbb80"} Dec 05 08:20:43 crc kubenswrapper[4876]: I1205 08:20:43.836663 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a785619-6b17-4496-bdaa-95244ef2bfb5" path="/var/lib/kubelet/pods/5a785619-6b17-4496-bdaa-95244ef2bfb5/volumes" Dec 05 08:20:44 crc kubenswrapper[4876]: I1205 08:20:44.324751 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" event={"ID":"68559e3a-898b-48f2-8a86-362dcfe74e78","Type":"ContainerStarted","Data":"002b4691965d15a5f6d4b56a3781a43992bd96b0bcd9b25c6984cfb03622f6b2"} Dec 05 08:20:44 crc kubenswrapper[4876]: I1205 08:20:44.326363 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:44 crc kubenswrapper[4876]: I1205 08:20:44.350622 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" podStartSLOduration=3.350590768 podStartE2EDuration="3.350590768s" podCreationTimestamp="2025-12-05 08:20:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:20:44.347036618 +0000 UTC m=+308.835701240" watchObservedRunningTime="2025-12-05 08:20:44.350590768 +0000 UTC m=+308.839255420" Dec 05 08:20:44 crc kubenswrapper[4876]: I1205 08:20:44.434564 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:20:51 crc kubenswrapper[4876]: I1205 08:20:51.903736 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-467hq"] Dec 05 08:20:51 crc kubenswrapper[4876]: I1205 08:20:51.905127 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:51 crc kubenswrapper[4876]: I1205 08:20:51.914377 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-467hq"] Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.020664 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6face095-3db3-45b1-8f91-e210ea77520a-registry-certificates\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.020729 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.020762 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6face095-3db3-45b1-8f91-e210ea77520a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.020790 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6face095-3db3-45b1-8f91-e210ea77520a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.020832 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5mbg\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-kube-api-access-j5mbg\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.020997 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-bound-sa-token\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.021068 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-registry-tls\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.021210 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6face095-3db3-45b1-8f91-e210ea77520a-trusted-ca\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.053693 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.123026 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6face095-3db3-45b1-8f91-e210ea77520a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.123218 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6face095-3db3-45b1-8f91-e210ea77520a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.123306 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5mbg\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-kube-api-access-j5mbg\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.123423 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-bound-sa-token\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.123535 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-registry-tls\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.123713 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6face095-3db3-45b1-8f91-e210ea77520a-trusted-ca\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.123768 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6face095-3db3-45b1-8f91-e210ea77520a-registry-certificates\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.124217 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6face095-3db3-45b1-8f91-e210ea77520a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.125511 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6face095-3db3-45b1-8f91-e210ea77520a-trusted-ca\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.128500 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6face095-3db3-45b1-8f91-e210ea77520a-registry-certificates\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.136207 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-registry-tls\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.136270 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6face095-3db3-45b1-8f91-e210ea77520a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.144401 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-bound-sa-token\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.148663 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5mbg\" (UniqueName: \"kubernetes.io/projected/6face095-3db3-45b1-8f91-e210ea77520a-kube-api-access-j5mbg\") pod \"image-registry-66df7c8f76-467hq\" (UID: \"6face095-3db3-45b1-8f91-e210ea77520a\") " pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.225835 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:52 crc kubenswrapper[4876]: I1205 08:20:52.675229 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-467hq"] Dec 05 08:20:52 crc kubenswrapper[4876]: W1205 08:20:52.686770 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6face095_3db3_45b1_8f91_e210ea77520a.slice/crio-159ce3e2eb053c47f69ce18e0d3460e22bbbfaae640d4ad3f808f2444e88c196 WatchSource:0}: Error finding container 159ce3e2eb053c47f69ce18e0d3460e22bbbfaae640d4ad3f808f2444e88c196: Status 404 returned error can't find the container with id 159ce3e2eb053c47f69ce18e0d3460e22bbbfaae640d4ad3f808f2444e88c196 Dec 05 08:20:53 crc kubenswrapper[4876]: I1205 08:20:53.374857 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-467hq" event={"ID":"6face095-3db3-45b1-8f91-e210ea77520a","Type":"ContainerStarted","Data":"3281f1879e178be9245dee4d626b25ea4e50a426787867ceff8d3557e7ed39a6"} Dec 05 08:20:53 crc kubenswrapper[4876]: I1205 08:20:53.375191 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-467hq" event={"ID":"6face095-3db3-45b1-8f91-e210ea77520a","Type":"ContainerStarted","Data":"159ce3e2eb053c47f69ce18e0d3460e22bbbfaae640d4ad3f808f2444e88c196"} Dec 05 08:20:53 crc kubenswrapper[4876]: I1205 08:20:53.375211 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:20:53 crc kubenswrapper[4876]: I1205 08:20:53.391812 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-467hq" podStartSLOduration=2.391735181 podStartE2EDuration="2.391735181s" podCreationTimestamp="2025-12-05 08:20:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:20:53.389272881 +0000 UTC m=+317.877937533" watchObservedRunningTime="2025-12-05 08:20:53.391735181 +0000 UTC m=+317.880399853" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.487615 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nnc6w"] Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.488590 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nnc6w" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="registry-server" containerID="cri-o://750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142" gracePeriod=30 Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.495478 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kdhx2"] Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.496231 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kdhx2" podUID="dfa346f8-9447-4712-b791-21bf718356b3" containerName="registry-server" containerID="cri-o://a4a5bd99215d91c38e65ded9d389c4f34f800fef598b9d0019f449c6d2362c5e" gracePeriod=30 Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.509122 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gj226"] Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.509312 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" containerID="cri-o://bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600" gracePeriod=30 Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.529840 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4g9z7"] Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.530135 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4g9z7" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerName="registry-server" containerID="cri-o://f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c" gracePeriod=30 Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.538157 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rf92g"] Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.538428 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rf92g" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="registry-server" containerID="cri-o://f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5" gracePeriod=30 Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.547746 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5hd7v"] Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.561225 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5hd7v"] Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.561328 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.619565 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv9p5\" (UniqueName: \"kubernetes.io/projected/e0282181-9f04-4c52-882f-322cd5a3ab11-kube-api-access-hv9p5\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.619665 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0282181-9f04-4c52-882f-322cd5a3ab11-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.619716 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e0282181-9f04-4c52-882f-322cd5a3ab11-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.720470 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0282181-9f04-4c52-882f-322cd5a3ab11-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.720542 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e0282181-9f04-4c52-882f-322cd5a3ab11-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.720582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv9p5\" (UniqueName: \"kubernetes.io/projected/e0282181-9f04-4c52-882f-322cd5a3ab11-kube-api-access-hv9p5\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.721968 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e0282181-9f04-4c52-882f-322cd5a3ab11-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.726867 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e0282181-9f04-4c52-882f-322cd5a3ab11-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:58 crc kubenswrapper[4876]: I1205 08:20:58.738420 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv9p5\" (UniqueName: \"kubernetes.io/projected/e0282181-9f04-4c52-882f-322cd5a3ab11-kube-api-access-hv9p5\") pod \"marketplace-operator-79b997595-5hd7v\" (UID: \"e0282181-9f04-4c52-882f-322cd5a3ab11\") " pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.044457 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.047710 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.128308 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-operator-metrics\") pod \"36087227-ee02-4dd0-b801-2ced90e4173f\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.128478 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zxbc\" (UniqueName: \"kubernetes.io/projected/36087227-ee02-4dd0-b801-2ced90e4173f-kube-api-access-5zxbc\") pod \"36087227-ee02-4dd0-b801-2ced90e4173f\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.128513 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-trusted-ca\") pod \"36087227-ee02-4dd0-b801-2ced90e4173f\" (UID: \"36087227-ee02-4dd0-b801-2ced90e4173f\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.129984 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "36087227-ee02-4dd0-b801-2ced90e4173f" (UID: "36087227-ee02-4dd0-b801-2ced90e4173f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.137790 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.151763 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "36087227-ee02-4dd0-b801-2ced90e4173f" (UID: "36087227-ee02-4dd0-b801-2ced90e4173f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.154478 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36087227-ee02-4dd0-b801-2ced90e4173f-kube-api-access-5zxbc" (OuterVolumeSpecName: "kube-api-access-5zxbc") pod "36087227-ee02-4dd0-b801-2ced90e4173f" (UID: "36087227-ee02-4dd0-b801-2ced90e4173f"). InnerVolumeSpecName "kube-api-access-5zxbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.165223 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.204205 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.234387 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zxbc\" (UniqueName: \"kubernetes.io/projected/36087227-ee02-4dd0-b801-2ced90e4173f-kube-api-access-5zxbc\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.234422 4876 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.234434 4876 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/36087227-ee02-4dd0-b801-2ced90e4173f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335140 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-utilities\") pod \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335206 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-utilities\") pod \"17ce8bf7-9544-4849-aa82-c5744a929656\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335257 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-catalog-content\") pod \"17ce8bf7-9544-4849-aa82-c5744a929656\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335279 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-catalog-content\") pod \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335341 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt7qv\" (UniqueName: \"kubernetes.io/projected/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-kube-api-access-zt7qv\") pod \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335362 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnqdv\" (UniqueName: \"kubernetes.io/projected/b48955db-dd1c-4e11-8d01-d465e6e7a75d-kube-api-access-cnqdv\") pod \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335403 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-catalog-content\") pod \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\" (UID: \"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335423 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrspc\" (UniqueName: \"kubernetes.io/projected/17ce8bf7-9544-4849-aa82-c5744a929656-kube-api-access-mrspc\") pod \"17ce8bf7-9544-4849-aa82-c5744a929656\" (UID: \"17ce8bf7-9544-4849-aa82-c5744a929656\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335437 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-utilities\") pod \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\" (UID: \"b48955db-dd1c-4e11-8d01-d465e6e7a75d\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.335837 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-utilities" (OuterVolumeSpecName: "utilities") pod "5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" (UID: "5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.336571 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-utilities" (OuterVolumeSpecName: "utilities") pod "b48955db-dd1c-4e11-8d01-d465e6e7a75d" (UID: "b48955db-dd1c-4e11-8d01-d465e6e7a75d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.337373 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-utilities" (OuterVolumeSpecName: "utilities") pod "17ce8bf7-9544-4849-aa82-c5744a929656" (UID: "17ce8bf7-9544-4849-aa82-c5744a929656"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.349599 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48955db-dd1c-4e11-8d01-d465e6e7a75d-kube-api-access-cnqdv" (OuterVolumeSpecName: "kube-api-access-cnqdv") pod "b48955db-dd1c-4e11-8d01-d465e6e7a75d" (UID: "b48955db-dd1c-4e11-8d01-d465e6e7a75d"). InnerVolumeSpecName "kube-api-access-cnqdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.350128 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17ce8bf7-9544-4849-aa82-c5744a929656-kube-api-access-mrspc" (OuterVolumeSpecName: "kube-api-access-mrspc") pod "17ce8bf7-9544-4849-aa82-c5744a929656" (UID: "17ce8bf7-9544-4849-aa82-c5744a929656"). InnerVolumeSpecName "kube-api-access-mrspc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.357714 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" (UID: "5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.358689 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-kube-api-access-zt7qv" (OuterVolumeSpecName: "kube-api-access-zt7qv") pod "5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" (UID: "5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e"). InnerVolumeSpecName "kube-api-access-zt7qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.395708 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17ce8bf7-9544-4849-aa82-c5744a929656" (UID: "17ce8bf7-9544-4849-aa82-c5744a929656"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.422056 4876 generic.go:334] "Generic (PLEG): container finished" podID="17ce8bf7-9544-4849-aa82-c5744a929656" containerID="750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142" exitCode=0 Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.422124 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnc6w" event={"ID":"17ce8bf7-9544-4849-aa82-c5744a929656","Type":"ContainerDied","Data":"750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.422153 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nnc6w" event={"ID":"17ce8bf7-9544-4849-aa82-c5744a929656","Type":"ContainerDied","Data":"ea1260633b2bca9f2eb6a220eb89ad59090e33b1cd3952ae29841fdd8f5a5cc2"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.422172 4876 scope.go:117] "RemoveContainer" containerID="750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.422170 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nnc6w" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.434495 4876 generic.go:334] "Generic (PLEG): container finished" podID="dfa346f8-9447-4712-b791-21bf718356b3" containerID="a4a5bd99215d91c38e65ded9d389c4f34f800fef598b9d0019f449c6d2362c5e" exitCode=0 Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.434573 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdhx2" event={"ID":"dfa346f8-9447-4712-b791-21bf718356b3","Type":"ContainerDied","Data":"a4a5bd99215d91c38e65ded9d389c4f34f800fef598b9d0019f449c6d2362c5e"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.437180 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnqdv\" (UniqueName: \"kubernetes.io/projected/b48955db-dd1c-4e11-8d01-d465e6e7a75d-kube-api-access-cnqdv\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.437209 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.437218 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrspc\" (UniqueName: \"kubernetes.io/projected/17ce8bf7-9544-4849-aa82-c5744a929656-kube-api-access-mrspc\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.437228 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.437238 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.437245 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.437253 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ce8bf7-9544-4849-aa82-c5744a929656-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.437261 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt7qv\" (UniqueName: \"kubernetes.io/projected/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e-kube-api-access-zt7qv\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.438450 4876 generic.go:334] "Generic (PLEG): container finished" podID="36087227-ee02-4dd0-b801-2ced90e4173f" containerID="bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600" exitCode=0 Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.438500 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" event={"ID":"36087227-ee02-4dd0-b801-2ced90e4173f","Type":"ContainerDied","Data":"bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.438523 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" event={"ID":"36087227-ee02-4dd0-b801-2ced90e4173f","Type":"ContainerDied","Data":"afe8123cbc3997d749110d179145b97f73c22aa4e993877ef7cc44e202248aeb"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.438606 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gj226" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.455086 4876 scope.go:117] "RemoveContainer" containerID="93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.462005 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b48955db-dd1c-4e11-8d01-d465e6e7a75d" (UID: "b48955db-dd1c-4e11-8d01-d465e6e7a75d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.463350 4876 generic.go:334] "Generic (PLEG): container finished" podID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerID="f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5" exitCode=0 Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.463686 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf92g" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.463835 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf92g" event={"ID":"b48955db-dd1c-4e11-8d01-d465e6e7a75d","Type":"ContainerDied","Data":"f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.464017 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf92g" event={"ID":"b48955db-dd1c-4e11-8d01-d465e6e7a75d","Type":"ContainerDied","Data":"85ea3158416128bf6b60af1030d1941902ca7d437b933fef45117b441e9c6028"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.482322 4876 generic.go:334] "Generic (PLEG): container finished" podID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerID="f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c" exitCode=0 Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.482373 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4g9z7" event={"ID":"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e","Type":"ContainerDied","Data":"f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.482406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4g9z7" event={"ID":"5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e","Type":"ContainerDied","Data":"6d32599322287784390563ecbfa4cc0bc60ce1e10a274c52e19533e2d5bf05b8"} Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.482473 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4g9z7" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.486537 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5hd7v"] Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.488925 4876 scope.go:117] "RemoveContainer" containerID="a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.506564 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gj226"] Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.509790 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gj226"] Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.525857 4876 scope.go:117] "RemoveContainer" containerID="750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.526461 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142\": container with ID starting with 750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142 not found: ID does not exist" containerID="750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.526505 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142"} err="failed to get container status \"750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142\": rpc error: code = NotFound desc = could not find container \"750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142\": container with ID starting with 750f9b17fe31781949cc415d160dd41f090d559ff2d791167d545fea187de142 not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.526530 4876 scope.go:117] "RemoveContainer" containerID="93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.526824 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2\": container with ID starting with 93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2 not found: ID does not exist" containerID="93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.526855 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2"} err="failed to get container status \"93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2\": rpc error: code = NotFound desc = could not find container \"93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2\": container with ID starting with 93bd8a3bafa11ef3cbed3f85598d03b2f9108ad0e62aef1e8308e73ed47979f2 not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.526878 4876 scope.go:117] "RemoveContainer" containerID="a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.527301 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327\": container with ID starting with a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327 not found: ID does not exist" containerID="a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.527326 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327"} err="failed to get container status \"a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327\": rpc error: code = NotFound desc = could not find container \"a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327\": container with ID starting with a4bbd006d5146487dc562b608417a254c119a81320aba17e3d592f255fc21327 not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.527342 4876 scope.go:117] "RemoveContainer" containerID="bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.532322 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.538188 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48955db-dd1c-4e11-8d01-d465e6e7a75d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.546320 4876 scope.go:117] "RemoveContainer" containerID="bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.561372 4876 scope.go:117] "RemoveContainer" containerID="bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.565226 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nnc6w"] Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.567237 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600\": container with ID starting with bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600 not found: ID does not exist" containerID="bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.567276 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600"} err="failed to get container status \"bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600\": rpc error: code = NotFound desc = could not find container \"bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600\": container with ID starting with bd9ecd643be5bf4dc026f9692b4963057a880473af88907855ab4e453a52c600 not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.567301 4876 scope.go:117] "RemoveContainer" containerID="bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.568510 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e\": container with ID starting with bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e not found: ID does not exist" containerID="bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.568527 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e"} err="failed to get container status \"bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e\": rpc error: code = NotFound desc = could not find container \"bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e\": container with ID starting with bf968054ab580a8d52e400e83faab9eae76526a24c583ec1702f76834919603e not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.568541 4876 scope.go:117] "RemoveContainer" containerID="f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.580226 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nnc6w"] Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.580283 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rf92g"] Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.580296 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rf92g"] Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.584748 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4g9z7"] Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.587928 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4g9z7"] Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.597867 4876 scope.go:117] "RemoveContainer" containerID="096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.616358 4876 scope.go:117] "RemoveContainer" containerID="e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.636737 4876 scope.go:117] "RemoveContainer" containerID="f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.637194 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5\": container with ID starting with f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5 not found: ID does not exist" containerID="f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.637228 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5"} err="failed to get container status \"f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5\": rpc error: code = NotFound desc = could not find container \"f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5\": container with ID starting with f0d0fefba531c8feac2368e781c5b7010f3715174b9a84abf12c5a153d8ceee5 not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.637254 4876 scope.go:117] "RemoveContainer" containerID="096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.637509 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e\": container with ID starting with 096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e not found: ID does not exist" containerID="096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.637532 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e"} err="failed to get container status \"096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e\": rpc error: code = NotFound desc = could not find container \"096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e\": container with ID starting with 096f205c65974b8c5808f60a0eed41dd03ff28fe9bcabfb52a06a2c27fe72b7e not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.637552 4876 scope.go:117] "RemoveContainer" containerID="e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.637873 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99\": container with ID starting with e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99 not found: ID does not exist" containerID="e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.637927 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99"} err="failed to get container status \"e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99\": rpc error: code = NotFound desc = could not find container \"e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99\": container with ID starting with e581b28070cf1d58225da6a9fe14e4b1c877df310d3df7c709c3acebe6ce2f99 not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.637956 4876 scope.go:117] "RemoveContainer" containerID="f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.638574 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-utilities\") pod \"dfa346f8-9447-4712-b791-21bf718356b3\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.638609 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-catalog-content\") pod \"dfa346f8-9447-4712-b791-21bf718356b3\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.638694 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdlbc\" (UniqueName: \"kubernetes.io/projected/dfa346f8-9447-4712-b791-21bf718356b3-kube-api-access-kdlbc\") pod \"dfa346f8-9447-4712-b791-21bf718356b3\" (UID: \"dfa346f8-9447-4712-b791-21bf718356b3\") " Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.639635 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-utilities" (OuterVolumeSpecName: "utilities") pod "dfa346f8-9447-4712-b791-21bf718356b3" (UID: "dfa346f8-9447-4712-b791-21bf718356b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.642165 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.648103 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfa346f8-9447-4712-b791-21bf718356b3-kube-api-access-kdlbc" (OuterVolumeSpecName: "kube-api-access-kdlbc") pod "dfa346f8-9447-4712-b791-21bf718356b3" (UID: "dfa346f8-9447-4712-b791-21bf718356b3"). InnerVolumeSpecName "kube-api-access-kdlbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.702465 4876 scope.go:117] "RemoveContainer" containerID="a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.710077 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfa346f8-9447-4712-b791-21bf718356b3" (UID: "dfa346f8-9447-4712-b791-21bf718356b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.715607 4876 scope.go:117] "RemoveContainer" containerID="877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.732452 4876 scope.go:117] "RemoveContainer" containerID="f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.733698 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c\": container with ID starting with f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c not found: ID does not exist" containerID="f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.733737 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c"} err="failed to get container status \"f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c\": rpc error: code = NotFound desc = could not find container \"f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c\": container with ID starting with f09873cb38b192bcdbce86f59daa46b4e3c550c8a9e384bf8bb288974f6b2b3c not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.733762 4876 scope.go:117] "RemoveContainer" containerID="a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.734855 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533\": container with ID starting with a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533 not found: ID does not exist" containerID="a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.734879 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533"} err="failed to get container status \"a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533\": rpc error: code = NotFound desc = could not find container \"a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533\": container with ID starting with a8d8c431f723889269589fe267c759fb6cff0d6695cd462411f8da125bb4e533 not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.734905 4876 scope.go:117] "RemoveContainer" containerID="877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f" Dec 05 08:20:59 crc kubenswrapper[4876]: E1205 08:20:59.735298 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f\": container with ID starting with 877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f not found: ID does not exist" containerID="877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.735320 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f"} err="failed to get container status \"877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f\": rpc error: code = NotFound desc = could not find container \"877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f\": container with ID starting with 877bb1df5bf5e6028ab3f6c2a260278df14cde482563bd641dde6e374d3d567f not found: ID does not exist" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.743274 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdlbc\" (UniqueName: \"kubernetes.io/projected/dfa346f8-9447-4712-b791-21bf718356b3-kube-api-access-kdlbc\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.743316 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfa346f8-9447-4712-b791-21bf718356b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.832546 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" path="/var/lib/kubelet/pods/17ce8bf7-9544-4849-aa82-c5744a929656/volumes" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.833540 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" path="/var/lib/kubelet/pods/36087227-ee02-4dd0-b801-2ced90e4173f/volumes" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.834175 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" path="/var/lib/kubelet/pods/5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e/volumes" Dec 05 08:20:59 crc kubenswrapper[4876]: I1205 08:20:59.835509 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" path="/var/lib/kubelet/pods/b48955db-dd1c-4e11-8d01-d465e6e7a75d/volumes" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.491460 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" event={"ID":"e0282181-9f04-4c52-882f-322cd5a3ab11","Type":"ContainerStarted","Data":"509d0de1ca276d5b4cd4025d7f4c30b15bca085fcb91bcb5a3e8575ca532363c"} Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.491500 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" event={"ID":"e0282181-9f04-4c52-882f-322cd5a3ab11","Type":"ContainerStarted","Data":"5bad25afaf129c106de8065939c52fa36ca5188fb1d345080ab51ced935e55c2"} Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.493251 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.499423 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kdhx2" event={"ID":"dfa346f8-9447-4712-b791-21bf718356b3","Type":"ContainerDied","Data":"5156aca98613496935a888e7333b99789c09ccfefb4c065b92110f0fc3d230c3"} Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.499477 4876 scope.go:117] "RemoveContainer" containerID="a4a5bd99215d91c38e65ded9d389c4f34f800fef598b9d0019f449c6d2362c5e" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.499477 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kdhx2" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.501130 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.511694 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5hd7v" podStartSLOduration=2.511678663 podStartE2EDuration="2.511678663s" podCreationTimestamp="2025-12-05 08:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:21:00.509927524 +0000 UTC m=+324.998592156" watchObservedRunningTime="2025-12-05 08:21:00.511678663 +0000 UTC m=+325.000343285" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.515817 4876 scope.go:117] "RemoveContainer" containerID="352c57fdbbfb1e20bdb97e12dd2aae7c1ae76d1c28e3f24cc34dee9e20bf0dd5" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.528729 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kdhx2"] Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.532224 4876 scope.go:117] "RemoveContainer" containerID="f8a94d614de8aa369b4405d9106cb2ec4e1c9183f5a20d42c74ee0b6998596e5" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.539464 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kdhx2"] Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701428 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bl5c6"] Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701639 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701653 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701666 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="extract-utilities" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701673 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="extract-utilities" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701683 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701692 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701702 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa346f8-9447-4712-b791-21bf718356b3" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701708 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa346f8-9447-4712-b791-21bf718356b3" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701718 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701725 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701733 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="extract-content" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701741 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="extract-content" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701751 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerName="extract-utilities" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701760 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerName="extract-utilities" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701769 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="extract-content" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701776 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="extract-content" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701788 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa346f8-9447-4712-b791-21bf718356b3" containerName="extract-utilities" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701796 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa346f8-9447-4712-b791-21bf718356b3" containerName="extract-utilities" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701806 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="extract-utilities" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701814 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="extract-utilities" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701823 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa346f8-9447-4712-b791-21bf718356b3" containerName="extract-content" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701830 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa346f8-9447-4712-b791-21bf718356b3" containerName="extract-content" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701846 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerName="extract-content" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701853 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerName="extract-content" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.701864 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701871 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.701995 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.702012 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ce8bf7-9544-4849-aa82-c5744a929656" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.702021 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa346f8-9447-4712-b791-21bf718356b3" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.702030 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48955db-dd1c-4e11-8d01-d465e6e7a75d" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.702058 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c1ee83e-b09e-4e3c-b533-6e1ad049eb8e" containerName="registry-server" Dec 05 08:21:00 crc kubenswrapper[4876]: E1205 08:21:00.702152 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.702162 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.702267 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="36087227-ee02-4dd0-b801-2ced90e4173f" containerName="marketplace-operator" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.702839 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.705486 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.716248 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bl5c6"] Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.858595 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b2nh\" (UniqueName: \"kubernetes.io/projected/2c4bad1b-8683-4357-a021-cdd84678aa19-kube-api-access-4b2nh\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.858650 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-catalog-content\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.858675 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-utilities\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.908961 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8n6mr"] Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.910806 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.913806 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.915617 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8n6mr"] Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.959577 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b2nh\" (UniqueName: \"kubernetes.io/projected/2c4bad1b-8683-4357-a021-cdd84678aa19-kube-api-access-4b2nh\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.959639 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-catalog-content\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.959664 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-utilities\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.960107 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-utilities\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.960365 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-catalog-content\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:00 crc kubenswrapper[4876]: I1205 08:21:00.980959 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b2nh\" (UniqueName: \"kubernetes.io/projected/2c4bad1b-8683-4357-a021-cdd84678aa19-kube-api-access-4b2nh\") pod \"certified-operators-bl5c6\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.018134 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.060781 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97e7ccac-0d35-45a2-a36c-974f008cf0cc-utilities\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.060971 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngtgm\" (UniqueName: \"kubernetes.io/projected/97e7ccac-0d35-45a2-a36c-974f008cf0cc-kube-api-access-ngtgm\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.061073 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97e7ccac-0d35-45a2-a36c-974f008cf0cc-catalog-content\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.162543 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97e7ccac-0d35-45a2-a36c-974f008cf0cc-utilities\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.162968 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngtgm\" (UniqueName: \"kubernetes.io/projected/97e7ccac-0d35-45a2-a36c-974f008cf0cc-kube-api-access-ngtgm\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.163023 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97e7ccac-0d35-45a2-a36c-974f008cf0cc-catalog-content\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.163551 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97e7ccac-0d35-45a2-a36c-974f008cf0cc-catalog-content\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.163584 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97e7ccac-0d35-45a2-a36c-974f008cf0cc-utilities\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.191792 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngtgm\" (UniqueName: \"kubernetes.io/projected/97e7ccac-0d35-45a2-a36c-974f008cf0cc-kube-api-access-ngtgm\") pod \"redhat-marketplace-8n6mr\" (UID: \"97e7ccac-0d35-45a2-a36c-974f008cf0cc\") " pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.226868 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.438857 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bl5c6"] Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.509334 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bl5c6" event={"ID":"2c4bad1b-8683-4357-a021-cdd84678aa19","Type":"ContainerStarted","Data":"dd9b95c8f0be41a43cbb084d0b0f980c76ee3cfba90eafc3661e1dd224d1daf3"} Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.602654 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8n6mr"] Dec 05 08:21:01 crc kubenswrapper[4876]: W1205 08:21:01.613105 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97e7ccac_0d35_45a2_a36c_974f008cf0cc.slice/crio-49c6701e23acedb428c665ceba6848227e6f2afd123ea2cd01e155a9e3ec0789 WatchSource:0}: Error finding container 49c6701e23acedb428c665ceba6848227e6f2afd123ea2cd01e155a9e3ec0789: Status 404 returned error can't find the container with id 49c6701e23acedb428c665ceba6848227e6f2afd123ea2cd01e155a9e3ec0789 Dec 05 08:21:01 crc kubenswrapper[4876]: I1205 08:21:01.838940 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfa346f8-9447-4712-b791-21bf718356b3" path="/var/lib/kubelet/pods/dfa346f8-9447-4712-b791-21bf718356b3/volumes" Dec 05 08:21:02 crc kubenswrapper[4876]: I1205 08:21:02.515811 4876 generic.go:334] "Generic (PLEG): container finished" podID="97e7ccac-0d35-45a2-a36c-974f008cf0cc" containerID="4f215191d7d177072f9cda57f428215ea57129996d0d9945115e8888343f6ac8" exitCode=0 Dec 05 08:21:02 crc kubenswrapper[4876]: I1205 08:21:02.515923 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n6mr" event={"ID":"97e7ccac-0d35-45a2-a36c-974f008cf0cc","Type":"ContainerDied","Data":"4f215191d7d177072f9cda57f428215ea57129996d0d9945115e8888343f6ac8"} Dec 05 08:21:02 crc kubenswrapper[4876]: I1205 08:21:02.516236 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n6mr" event={"ID":"97e7ccac-0d35-45a2-a36c-974f008cf0cc","Type":"ContainerStarted","Data":"49c6701e23acedb428c665ceba6848227e6f2afd123ea2cd01e155a9e3ec0789"} Dec 05 08:21:02 crc kubenswrapper[4876]: I1205 08:21:02.520111 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bl5c6" event={"ID":"2c4bad1b-8683-4357-a021-cdd84678aa19","Type":"ContainerDied","Data":"8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b"} Dec 05 08:21:02 crc kubenswrapper[4876]: I1205 08:21:02.525252 4876 generic.go:334] "Generic (PLEG): container finished" podID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerID="8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b" exitCode=0 Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.101251 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rlwz2"] Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.102477 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.106162 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.109266 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rlwz2"] Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.295730 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-catalog-content\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.295806 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sqs5\" (UniqueName: \"kubernetes.io/projected/c58822ac-9c07-4118-8743-8ace4b299aa1-kube-api-access-6sqs5\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.295844 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-utilities\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.303590 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-89hsw"] Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.304809 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.307736 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.308699 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-89hsw"] Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.397413 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-catalog-content\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.397723 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xnpb\" (UniqueName: \"kubernetes.io/projected/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-kube-api-access-5xnpb\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.397884 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-utilities\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.398015 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sqs5\" (UniqueName: \"kubernetes.io/projected/c58822ac-9c07-4118-8743-8ace4b299aa1-kube-api-access-6sqs5\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.398391 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-utilities\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.398838 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-catalog-content\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.398740 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-utilities\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.398290 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-catalog-content\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.421719 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sqs5\" (UniqueName: \"kubernetes.io/projected/c58822ac-9c07-4118-8743-8ace4b299aa1-kube-api-access-6sqs5\") pod \"community-operators-rlwz2\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.455055 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.500410 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-utilities\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.500503 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-catalog-content\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.500582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xnpb\" (UniqueName: \"kubernetes.io/projected/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-kube-api-access-5xnpb\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.501179 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-utilities\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.501509 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-catalog-content\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.528950 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xnpb\" (UniqueName: \"kubernetes.io/projected/3df610b8-24ff-45eb-92cd-8fd40cd53f0c-kube-api-access-5xnpb\") pod \"redhat-operators-89hsw\" (UID: \"3df610b8-24ff-45eb-92cd-8fd40cd53f0c\") " pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.534436 4876 generic.go:334] "Generic (PLEG): container finished" podID="97e7ccac-0d35-45a2-a36c-974f008cf0cc" containerID="d8a20d1158224ef822c0423a756840819689d490723c7c16ecbda46698901301" exitCode=0 Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.534521 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n6mr" event={"ID":"97e7ccac-0d35-45a2-a36c-974f008cf0cc","Type":"ContainerDied","Data":"d8a20d1158224ef822c0423a756840819689d490723c7c16ecbda46698901301"} Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.546565 4876 generic.go:334] "Generic (PLEG): container finished" podID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerID="2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c" exitCode=0 Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.546623 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bl5c6" event={"ID":"2c4bad1b-8683-4357-a021-cdd84678aa19","Type":"ContainerDied","Data":"2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c"} Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.624977 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:03 crc kubenswrapper[4876]: I1205 08:21:03.876091 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rlwz2"] Dec 05 08:21:03 crc kubenswrapper[4876]: W1205 08:21:03.876542 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc58822ac_9c07_4118_8743_8ace4b299aa1.slice/crio-96b09de799f5d821590d5b4b2213bf717b3f9ec027af6c3f7b26d0d95e4926d9 WatchSource:0}: Error finding container 96b09de799f5d821590d5b4b2213bf717b3f9ec027af6c3f7b26d0d95e4926d9: Status 404 returned error can't find the container with id 96b09de799f5d821590d5b4b2213bf717b3f9ec027af6c3f7b26d0d95e4926d9 Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.040233 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-89hsw"] Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.553937 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bl5c6" event={"ID":"2c4bad1b-8683-4357-a021-cdd84678aa19","Type":"ContainerStarted","Data":"7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6"} Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.555703 4876 generic.go:334] "Generic (PLEG): container finished" podID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerID="0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c" exitCode=0 Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.555784 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlwz2" event={"ID":"c58822ac-9c07-4118-8743-8ace4b299aa1","Type":"ContainerDied","Data":"0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c"} Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.555815 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlwz2" event={"ID":"c58822ac-9c07-4118-8743-8ace4b299aa1","Type":"ContainerStarted","Data":"96b09de799f5d821590d5b4b2213bf717b3f9ec027af6c3f7b26d0d95e4926d9"} Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.556941 4876 generic.go:334] "Generic (PLEG): container finished" podID="3df610b8-24ff-45eb-92cd-8fd40cd53f0c" containerID="95bf9775c95117d00772b3ce4dcd9fd29c0c7645863e20b180705eca4e9d30fe" exitCode=0 Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.557037 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89hsw" event={"ID":"3df610b8-24ff-45eb-92cd-8fd40cd53f0c","Type":"ContainerDied","Data":"95bf9775c95117d00772b3ce4dcd9fd29c0c7645863e20b180705eca4e9d30fe"} Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.557074 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89hsw" event={"ID":"3df610b8-24ff-45eb-92cd-8fd40cd53f0c","Type":"ContainerStarted","Data":"d6d78d14dff26df4d6003e4d6e305a36c292a5b7ecb4bcf37a2e544be06ade5c"} Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.559525 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8n6mr" event={"ID":"97e7ccac-0d35-45a2-a36c-974f008cf0cc","Type":"ContainerStarted","Data":"85dd376693dbba5f0bf7a35fe0bb6df05573b3a86636918c1317238b6135af77"} Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.596740 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bl5c6" podStartSLOduration=3.183014817 podStartE2EDuration="4.596724189s" podCreationTimestamp="2025-12-05 08:21:00 +0000 UTC" firstStartedPulling="2025-12-05 08:21:02.526588953 +0000 UTC m=+327.015253615" lastFinishedPulling="2025-12-05 08:21:03.940298365 +0000 UTC m=+328.428962987" observedRunningTime="2025-12-05 08:21:04.574239274 +0000 UTC m=+329.062903916" watchObservedRunningTime="2025-12-05 08:21:04.596724189 +0000 UTC m=+329.085388801" Dec 05 08:21:04 crc kubenswrapper[4876]: I1205 08:21:04.601181 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8n6mr" podStartSLOduration=3.123303718 podStartE2EDuration="4.601171842s" podCreationTimestamp="2025-12-05 08:21:00 +0000 UTC" firstStartedPulling="2025-12-05 08:21:02.518113058 +0000 UTC m=+327.006777680" lastFinishedPulling="2025-12-05 08:21:03.995981182 +0000 UTC m=+328.484645804" observedRunningTime="2025-12-05 08:21:04.595794863 +0000 UTC m=+329.084459495" watchObservedRunningTime="2025-12-05 08:21:04.601171842 +0000 UTC m=+329.089836464" Dec 05 08:21:05 crc kubenswrapper[4876]: I1205 08:21:05.567540 4876 generic.go:334] "Generic (PLEG): container finished" podID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerID="6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4" exitCode=0 Dec 05 08:21:05 crc kubenswrapper[4876]: I1205 08:21:05.567805 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlwz2" event={"ID":"c58822ac-9c07-4118-8743-8ace4b299aa1","Type":"ContainerDied","Data":"6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4"} Dec 05 08:21:05 crc kubenswrapper[4876]: I1205 08:21:05.582893 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89hsw" event={"ID":"3df610b8-24ff-45eb-92cd-8fd40cd53f0c","Type":"ContainerStarted","Data":"d7b0c3bdc130b509fd686e08da099bd749ef4dacfbdda08ca712170d89ec5357"} Dec 05 08:21:06 crc kubenswrapper[4876]: I1205 08:21:06.591170 4876 generic.go:334] "Generic (PLEG): container finished" podID="3df610b8-24ff-45eb-92cd-8fd40cd53f0c" containerID="d7b0c3bdc130b509fd686e08da099bd749ef4dacfbdda08ca712170d89ec5357" exitCode=0 Dec 05 08:21:06 crc kubenswrapper[4876]: I1205 08:21:06.591228 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89hsw" event={"ID":"3df610b8-24ff-45eb-92cd-8fd40cd53f0c","Type":"ContainerDied","Data":"d7b0c3bdc130b509fd686e08da099bd749ef4dacfbdda08ca712170d89ec5357"} Dec 05 08:21:06 crc kubenswrapper[4876]: I1205 08:21:06.594621 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlwz2" event={"ID":"c58822ac-9c07-4118-8743-8ace4b299aa1","Type":"ContainerStarted","Data":"3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb"} Dec 05 08:21:06 crc kubenswrapper[4876]: I1205 08:21:06.635048 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rlwz2" podStartSLOduration=2.003373931 podStartE2EDuration="3.635032009s" podCreationTimestamp="2025-12-05 08:21:03 +0000 UTC" firstStartedPulling="2025-12-05 08:21:04.557470518 +0000 UTC m=+329.046135140" lastFinishedPulling="2025-12-05 08:21:06.189128596 +0000 UTC m=+330.677793218" observedRunningTime="2025-12-05 08:21:06.631273874 +0000 UTC m=+331.119938536" watchObservedRunningTime="2025-12-05 08:21:06.635032009 +0000 UTC m=+331.123696631" Dec 05 08:21:07 crc kubenswrapper[4876]: I1205 08:21:07.601046 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89hsw" event={"ID":"3df610b8-24ff-45eb-92cd-8fd40cd53f0c","Type":"ContainerStarted","Data":"4cbed00ddd674f8d77c69f68b309158c6f4cfcd8500ac2b589099d8762c53773"} Dec 05 08:21:07 crc kubenswrapper[4876]: I1205 08:21:07.620496 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-89hsw" podStartSLOduration=2.189072691 podStartE2EDuration="4.620479627s" podCreationTimestamp="2025-12-05 08:21:03 +0000 UTC" firstStartedPulling="2025-12-05 08:21:04.557782676 +0000 UTC m=+329.046447298" lastFinishedPulling="2025-12-05 08:21:06.989189572 +0000 UTC m=+331.477854234" observedRunningTime="2025-12-05 08:21:07.615998993 +0000 UTC m=+332.104663615" watchObservedRunningTime="2025-12-05 08:21:07.620479627 +0000 UTC m=+332.109144249" Dec 05 08:21:08 crc kubenswrapper[4876]: I1205 08:21:08.213716 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:21:08 crc kubenswrapper[4876]: I1205 08:21:08.214224 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:21:11 crc kubenswrapper[4876]: I1205 08:21:11.019129 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:11 crc kubenswrapper[4876]: I1205 08:21:11.019873 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:11 crc kubenswrapper[4876]: I1205 08:21:11.066664 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:11 crc kubenswrapper[4876]: I1205 08:21:11.227358 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:11 crc kubenswrapper[4876]: I1205 08:21:11.227790 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:11 crc kubenswrapper[4876]: I1205 08:21:11.278855 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:11 crc kubenswrapper[4876]: I1205 08:21:11.667261 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8n6mr" Dec 05 08:21:11 crc kubenswrapper[4876]: I1205 08:21:11.668889 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:21:12 crc kubenswrapper[4876]: I1205 08:21:12.232882 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-467hq" Dec 05 08:21:12 crc kubenswrapper[4876]: I1205 08:21:12.279646 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sklcd"] Dec 05 08:21:13 crc kubenswrapper[4876]: I1205 08:21:13.455504 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:13 crc kubenswrapper[4876]: I1205 08:21:13.456784 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:13 crc kubenswrapper[4876]: I1205 08:21:13.513199 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:13 crc kubenswrapper[4876]: I1205 08:21:13.626592 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:13 crc kubenswrapper[4876]: I1205 08:21:13.626719 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:13 crc kubenswrapper[4876]: I1205 08:21:13.675456 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:13 crc kubenswrapper[4876]: I1205 08:21:13.710129 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rlwz2" Dec 05 08:21:14 crc kubenswrapper[4876]: I1205 08:21:14.698315 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-89hsw" Dec 05 08:21:34 crc kubenswrapper[4876]: I1205 08:21:34.480749 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs"] Dec 05 08:21:34 crc kubenswrapper[4876]: I1205 08:21:34.481480 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" podUID="68559e3a-898b-48f2-8a86-362dcfe74e78" containerName="route-controller-manager" containerID="cri-o://002b4691965d15a5f6d4b56a3781a43992bd96b0bcd9b25c6984cfb03622f6b2" gracePeriod=30 Dec 05 08:21:34 crc kubenswrapper[4876]: I1205 08:21:34.738284 4876 generic.go:334] "Generic (PLEG): container finished" podID="68559e3a-898b-48f2-8a86-362dcfe74e78" containerID="002b4691965d15a5f6d4b56a3781a43992bd96b0bcd9b25c6984cfb03622f6b2" exitCode=0 Dec 05 08:21:34 crc kubenswrapper[4876]: I1205 08:21:34.738331 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" event={"ID":"68559e3a-898b-48f2-8a86-362dcfe74e78","Type":"ContainerDied","Data":"002b4691965d15a5f6d4b56a3781a43992bd96b0bcd9b25c6984cfb03622f6b2"} Dec 05 08:21:34 crc kubenswrapper[4876]: I1205 08:21:34.852548 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.024446 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-config\") pod \"68559e3a-898b-48f2-8a86-362dcfe74e78\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.024565 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-client-ca\") pod \"68559e3a-898b-48f2-8a86-362dcfe74e78\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.024697 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgf9m\" (UniqueName: \"kubernetes.io/projected/68559e3a-898b-48f2-8a86-362dcfe74e78-kube-api-access-xgf9m\") pod \"68559e3a-898b-48f2-8a86-362dcfe74e78\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.024754 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68559e3a-898b-48f2-8a86-362dcfe74e78-serving-cert\") pod \"68559e3a-898b-48f2-8a86-362dcfe74e78\" (UID: \"68559e3a-898b-48f2-8a86-362dcfe74e78\") " Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.025448 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-config" (OuterVolumeSpecName: "config") pod "68559e3a-898b-48f2-8a86-362dcfe74e78" (UID: "68559e3a-898b-48f2-8a86-362dcfe74e78"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.025789 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-client-ca" (OuterVolumeSpecName: "client-ca") pod "68559e3a-898b-48f2-8a86-362dcfe74e78" (UID: "68559e3a-898b-48f2-8a86-362dcfe74e78"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.030347 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68559e3a-898b-48f2-8a86-362dcfe74e78-kube-api-access-xgf9m" (OuterVolumeSpecName: "kube-api-access-xgf9m") pod "68559e3a-898b-48f2-8a86-362dcfe74e78" (UID: "68559e3a-898b-48f2-8a86-362dcfe74e78"). InnerVolumeSpecName "kube-api-access-xgf9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.039675 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68559e3a-898b-48f2-8a86-362dcfe74e78-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "68559e3a-898b-48f2-8a86-362dcfe74e78" (UID: "68559e3a-898b-48f2-8a86-362dcfe74e78"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.126957 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.127006 4876 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68559e3a-898b-48f2-8a86-362dcfe74e78-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.127026 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgf9m\" (UniqueName: \"kubernetes.io/projected/68559e3a-898b-48f2-8a86-362dcfe74e78-kube-api-access-xgf9m\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.127047 4876 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68559e3a-898b-48f2-8a86-362dcfe74e78-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.747141 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" event={"ID":"68559e3a-898b-48f2-8a86-362dcfe74e78","Type":"ContainerDied","Data":"da8fda97687c9e89bb776eff39c9082ff9e7eef5c2c9af7dbefab6d24c4bbb80"} Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.747151 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.747196 4876 scope.go:117] "RemoveContainer" containerID="002b4691965d15a5f6d4b56a3781a43992bd96b0bcd9b25c6984cfb03622f6b2" Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.792131 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs"] Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.795637 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8678c97bff-lhzhs"] Dec 05 08:21:35 crc kubenswrapper[4876]: I1205 08:21:35.835595 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68559e3a-898b-48f2-8a86-362dcfe74e78" path="/var/lib/kubelet/pods/68559e3a-898b-48f2-8a86-362dcfe74e78/volumes" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.516675 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh"] Dec 05 08:21:36 crc kubenswrapper[4876]: E1205 08:21:36.517609 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68559e3a-898b-48f2-8a86-362dcfe74e78" containerName="route-controller-manager" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.517723 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="68559e3a-898b-48f2-8a86-362dcfe74e78" containerName="route-controller-manager" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.517983 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="68559e3a-898b-48f2-8a86-362dcfe74e78" containerName="route-controller-manager" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.518545 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.522555 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.522632 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.523400 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.523831 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.524031 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.524511 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.530359 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh"] Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.647278 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75a32dc3-a32e-4589-91dc-81fde37ff1c8-serving-cert\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.647341 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75a32dc3-a32e-4589-91dc-81fde37ff1c8-client-ca\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.647366 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk89z\" (UniqueName: \"kubernetes.io/projected/75a32dc3-a32e-4589-91dc-81fde37ff1c8-kube-api-access-pk89z\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.647390 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75a32dc3-a32e-4589-91dc-81fde37ff1c8-config\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.748257 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75a32dc3-a32e-4589-91dc-81fde37ff1c8-client-ca\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.748307 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk89z\" (UniqueName: \"kubernetes.io/projected/75a32dc3-a32e-4589-91dc-81fde37ff1c8-kube-api-access-pk89z\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.748345 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75a32dc3-a32e-4589-91dc-81fde37ff1c8-config\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.748412 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75a32dc3-a32e-4589-91dc-81fde37ff1c8-serving-cert\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.750077 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75a32dc3-a32e-4589-91dc-81fde37ff1c8-config\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.750777 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/75a32dc3-a32e-4589-91dc-81fde37ff1c8-client-ca\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.755181 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75a32dc3-a32e-4589-91dc-81fde37ff1c8-serving-cert\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.765117 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk89z\" (UniqueName: \"kubernetes.io/projected/75a32dc3-a32e-4589-91dc-81fde37ff1c8-kube-api-access-pk89z\") pod \"route-controller-manager-697f44cb6c-mzjgh\" (UID: \"75a32dc3-a32e-4589-91dc-81fde37ff1c8\") " pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:36 crc kubenswrapper[4876]: I1205 08:21:36.835363 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.245527 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh"] Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.330561 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" podUID="907ad872-3805-403d-9649-11e7090bb789" containerName="registry" containerID="cri-o://e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76" gracePeriod=30 Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.721981 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.764845 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" event={"ID":"75a32dc3-a32e-4589-91dc-81fde37ff1c8","Type":"ContainerStarted","Data":"34b60695b9239dc476c77da3fade038dd802d1d09377a405dc105888228c3c89"} Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.764892 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" event={"ID":"75a32dc3-a32e-4589-91dc-81fde37ff1c8","Type":"ContainerStarted","Data":"aaff3091aa13343f6ba6f387b2935dd601a9378ece2f26bd14e159dff3c06d9e"} Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.765845 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.773579 4876 generic.go:334] "Generic (PLEG): container finished" podID="907ad872-3805-403d-9649-11e7090bb789" containerID="e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76" exitCode=0 Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.773621 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" event={"ID":"907ad872-3805-403d-9649-11e7090bb789","Type":"ContainerDied","Data":"e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76"} Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.773648 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" event={"ID":"907ad872-3805-403d-9649-11e7090bb789","Type":"ContainerDied","Data":"60ad0c1095f6e57afcb208b478242c59b62a66aab37c2e1b514bc9c0fe165fad"} Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.773671 4876 scope.go:117] "RemoveContainer" containerID="e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.773751 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-sklcd" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.791379 4876 scope.go:117] "RemoveContainer" containerID="e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76" Dec 05 08:21:37 crc kubenswrapper[4876]: E1205 08:21:37.795056 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76\": container with ID starting with e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76 not found: ID does not exist" containerID="e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.795101 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76"} err="failed to get container status \"e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76\": rpc error: code = NotFound desc = could not find container \"e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76\": container with ID starting with e8a0a6789427f360f9764145da71a59a0153a2035372057f6e68ea7c3cd1cd76 not found: ID does not exist" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.798257 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" podStartSLOduration=3.798241406 podStartE2EDuration="3.798241406s" podCreationTimestamp="2025-12-05 08:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:21:37.795240723 +0000 UTC m=+362.283905345" watchObservedRunningTime="2025-12-05 08:21:37.798241406 +0000 UTC m=+362.286906018" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.863160 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/907ad872-3805-403d-9649-11e7090bb789-installation-pull-secrets\") pod \"907ad872-3805-403d-9649-11e7090bb789\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.863243 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhm65\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-kube-api-access-xhm65\") pod \"907ad872-3805-403d-9649-11e7090bb789\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.863407 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"907ad872-3805-403d-9649-11e7090bb789\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.863446 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-bound-sa-token\") pod \"907ad872-3805-403d-9649-11e7090bb789\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.863469 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-registry-certificates\") pod \"907ad872-3805-403d-9649-11e7090bb789\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.863486 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/907ad872-3805-403d-9649-11e7090bb789-ca-trust-extracted\") pod \"907ad872-3805-403d-9649-11e7090bb789\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.863522 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-registry-tls\") pod \"907ad872-3805-403d-9649-11e7090bb789\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.863552 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-trusted-ca\") pod \"907ad872-3805-403d-9649-11e7090bb789\" (UID: \"907ad872-3805-403d-9649-11e7090bb789\") " Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.864386 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "907ad872-3805-403d-9649-11e7090bb789" (UID: "907ad872-3805-403d-9649-11e7090bb789"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.865157 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "907ad872-3805-403d-9649-11e7090bb789" (UID: "907ad872-3805-403d-9649-11e7090bb789"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.869448 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "907ad872-3805-403d-9649-11e7090bb789" (UID: "907ad872-3805-403d-9649-11e7090bb789"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.870229 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/907ad872-3805-403d-9649-11e7090bb789-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "907ad872-3805-403d-9649-11e7090bb789" (UID: "907ad872-3805-403d-9649-11e7090bb789"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.870469 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-kube-api-access-xhm65" (OuterVolumeSpecName: "kube-api-access-xhm65") pod "907ad872-3805-403d-9649-11e7090bb789" (UID: "907ad872-3805-403d-9649-11e7090bb789"). InnerVolumeSpecName "kube-api-access-xhm65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.873968 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "907ad872-3805-403d-9649-11e7090bb789" (UID: "907ad872-3805-403d-9649-11e7090bb789"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.882851 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "907ad872-3805-403d-9649-11e7090bb789" (UID: "907ad872-3805-403d-9649-11e7090bb789"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.889396 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/907ad872-3805-403d-9649-11e7090bb789-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "907ad872-3805-403d-9649-11e7090bb789" (UID: "907ad872-3805-403d-9649-11e7090bb789"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.966548 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.966882 4876 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/907ad872-3805-403d-9649-11e7090bb789-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.967000 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhm65\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-kube-api-access-xhm65\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.967085 4876 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.967182 4876 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/907ad872-3805-403d-9649-11e7090bb789-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.967287 4876 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/907ad872-3805-403d-9649-11e7090bb789-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.968022 4876 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/907ad872-3805-403d-9649-11e7090bb789-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:21:37 crc kubenswrapper[4876]: I1205 08:21:37.995582 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-697f44cb6c-mzjgh" Dec 05 08:21:38 crc kubenswrapper[4876]: I1205 08:21:38.099363 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sklcd"] Dec 05 08:21:38 crc kubenswrapper[4876]: I1205 08:21:38.107202 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-sklcd"] Dec 05 08:21:38 crc kubenswrapper[4876]: I1205 08:21:38.213255 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:21:38 crc kubenswrapper[4876]: I1205 08:21:38.213302 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:21:39 crc kubenswrapper[4876]: I1205 08:21:39.832712 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="907ad872-3805-403d-9649-11e7090bb789" path="/var/lib/kubelet/pods/907ad872-3805-403d-9649-11e7090bb789/volumes" Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.213664 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.214432 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.214503 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.216350 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92b82d84a70d1805e62d229730a17c91d2767035530b63f457f1a8e1f0171d73"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.216493 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://92b82d84a70d1805e62d229730a17c91d2767035530b63f457f1a8e1f0171d73" gracePeriod=600 Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.968559 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="92b82d84a70d1805e62d229730a17c91d2767035530b63f457f1a8e1f0171d73" exitCode=0 Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.968596 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"92b82d84a70d1805e62d229730a17c91d2767035530b63f457f1a8e1f0171d73"} Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.968856 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"d39a251bae34b6211c7b1fa50cc0f96420717f4b2425b92d21b2365fcb057a2f"} Dec 05 08:22:08 crc kubenswrapper[4876]: I1205 08:22:08.968873 4876 scope.go:117] "RemoveContainer" containerID="2785fd50b41e6a50917fbec03f6da0ee28efef9c1fd814d437c74e1d957854ac" Dec 05 08:24:08 crc kubenswrapper[4876]: I1205 08:24:08.214036 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:24:08 crc kubenswrapper[4876]: I1205 08:24:08.214659 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:24:36 crc kubenswrapper[4876]: I1205 08:24:36.063416 4876 scope.go:117] "RemoveContainer" containerID="f93c474830bfcb8e4e41b0f1c09f83632caf1d5d2b13e906e487f85f7f3cde07" Dec 05 08:24:38 crc kubenswrapper[4876]: I1205 08:24:38.214630 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:24:38 crc kubenswrapper[4876]: I1205 08:24:38.215072 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:25:08 crc kubenswrapper[4876]: I1205 08:25:08.213787 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:25:08 crc kubenswrapper[4876]: I1205 08:25:08.214765 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:25:08 crc kubenswrapper[4876]: I1205 08:25:08.214978 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:25:08 crc kubenswrapper[4876]: I1205 08:25:08.216513 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d39a251bae34b6211c7b1fa50cc0f96420717f4b2425b92d21b2365fcb057a2f"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:25:08 crc kubenswrapper[4876]: I1205 08:25:08.216711 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://d39a251bae34b6211c7b1fa50cc0f96420717f4b2425b92d21b2365fcb057a2f" gracePeriod=600 Dec 05 08:25:09 crc kubenswrapper[4876]: I1205 08:25:09.076695 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="d39a251bae34b6211c7b1fa50cc0f96420717f4b2425b92d21b2365fcb057a2f" exitCode=0 Dec 05 08:25:09 crc kubenswrapper[4876]: I1205 08:25:09.076771 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"d39a251bae34b6211c7b1fa50cc0f96420717f4b2425b92d21b2365fcb057a2f"} Dec 05 08:25:09 crc kubenswrapper[4876]: I1205 08:25:09.077229 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"3323fa5b6329c4296cfe91ae288ce25651562ac0fb3157978c65f5ccba6cd828"} Dec 05 08:25:09 crc kubenswrapper[4876]: I1205 08:25:09.077267 4876 scope.go:117] "RemoveContainer" containerID="92b82d84a70d1805e62d229730a17c91d2767035530b63f457f1a8e1f0171d73" Dec 05 08:25:36 crc kubenswrapper[4876]: I1205 08:25:36.099293 4876 scope.go:117] "RemoveContainer" containerID="a59b0d6fff5cfe773dc8af339b59e7d6ea1fa88b5dd56460df073d7b0e5559bd" Dec 05 08:25:36 crc kubenswrapper[4876]: I1205 08:25:36.150821 4876 scope.go:117] "RemoveContainer" containerID="94d777af08a96b450c3f4189af6a7e710a9875cba82a3893125592a897b420c5" Dec 05 08:27:08 crc kubenswrapper[4876]: I1205 08:27:08.213517 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:27:08 crc kubenswrapper[4876]: I1205 08:27:08.214063 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:27:38 crc kubenswrapper[4876]: I1205 08:27:38.214082 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:27:38 crc kubenswrapper[4876]: I1205 08:27:38.214663 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:28:07 crc kubenswrapper[4876]: I1205 08:28:07.012633 4876 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 08:28:08 crc kubenswrapper[4876]: I1205 08:28:08.213926 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:28:08 crc kubenswrapper[4876]: I1205 08:28:08.214008 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:28:08 crc kubenswrapper[4876]: I1205 08:28:08.214070 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:28:08 crc kubenswrapper[4876]: I1205 08:28:08.214933 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3323fa5b6329c4296cfe91ae288ce25651562ac0fb3157978c65f5ccba6cd828"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:28:08 crc kubenswrapper[4876]: I1205 08:28:08.215031 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://3323fa5b6329c4296cfe91ae288ce25651562ac0fb3157978c65f5ccba6cd828" gracePeriod=600 Dec 05 08:28:09 crc kubenswrapper[4876]: I1205 08:28:09.145355 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="3323fa5b6329c4296cfe91ae288ce25651562ac0fb3157978c65f5ccba6cd828" exitCode=0 Dec 05 08:28:09 crc kubenswrapper[4876]: I1205 08:28:09.145406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"3323fa5b6329c4296cfe91ae288ce25651562ac0fb3157978c65f5ccba6cd828"} Dec 05 08:28:09 crc kubenswrapper[4876]: I1205 08:28:09.146014 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"b93e463bd75bc474356eb877c940a3ca93130dc5e558c7d99bfe4997ecfbcedb"} Dec 05 08:28:09 crc kubenswrapper[4876]: I1205 08:28:09.146047 4876 scope.go:117] "RemoveContainer" containerID="d39a251bae34b6211c7b1fa50cc0f96420717f4b2425b92d21b2365fcb057a2f" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.430222 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9n88g"] Dec 05 08:28:12 crc kubenswrapper[4876]: E1205 08:28:12.430839 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="907ad872-3805-403d-9649-11e7090bb789" containerName="registry" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.430860 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="907ad872-3805-403d-9649-11e7090bb789" containerName="registry" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.431017 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="907ad872-3805-403d-9649-11e7090bb789" containerName="registry" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.432057 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.451487 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9n88g"] Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.592588 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w86hg\" (UniqueName: \"kubernetes.io/projected/2739f16b-aa7a-45fd-bee3-e69f221176fd-kube-api-access-w86hg\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.592759 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-catalog-content\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.592791 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-utilities\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.693456 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-utilities\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.693552 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w86hg\" (UniqueName: \"kubernetes.io/projected/2739f16b-aa7a-45fd-bee3-e69f221176fd-kube-api-access-w86hg\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.693634 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-catalog-content\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.694034 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-utilities\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.694060 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-catalog-content\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.720079 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w86hg\" (UniqueName: \"kubernetes.io/projected/2739f16b-aa7a-45fd-bee3-e69f221176fd-kube-api-access-w86hg\") pod \"community-operators-9n88g\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:12 crc kubenswrapper[4876]: I1205 08:28:12.799015 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:13 crc kubenswrapper[4876]: I1205 08:28:13.053203 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9n88g"] Dec 05 08:28:13 crc kubenswrapper[4876]: I1205 08:28:13.173230 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9n88g" event={"ID":"2739f16b-aa7a-45fd-bee3-e69f221176fd","Type":"ContainerStarted","Data":"aa830717b9e3cf4692f0e32ced5cf926bb07c5a4c81cd7c81d27b163f72df1b3"} Dec 05 08:28:14 crc kubenswrapper[4876]: I1205 08:28:14.181281 4876 generic.go:334] "Generic (PLEG): container finished" podID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerID="02aed35fcf66b7eb8133f3490b0d94a933aec962ff7c55cfa69619fa6b5a4c0f" exitCode=0 Dec 05 08:28:14 crc kubenswrapper[4876]: I1205 08:28:14.181361 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9n88g" event={"ID":"2739f16b-aa7a-45fd-bee3-e69f221176fd","Type":"ContainerDied","Data":"02aed35fcf66b7eb8133f3490b0d94a933aec962ff7c55cfa69619fa6b5a4c0f"} Dec 05 08:28:14 crc kubenswrapper[4876]: I1205 08:28:14.183722 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:28:15 crc kubenswrapper[4876]: I1205 08:28:15.190615 4876 generic.go:334] "Generic (PLEG): container finished" podID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerID="f439b42a17362e20dc03bf17e7a154888f5630d689d6f7251228c8a64b0c6fda" exitCode=0 Dec 05 08:28:15 crc kubenswrapper[4876]: I1205 08:28:15.190691 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9n88g" event={"ID":"2739f16b-aa7a-45fd-bee3-e69f221176fd","Type":"ContainerDied","Data":"f439b42a17362e20dc03bf17e7a154888f5630d689d6f7251228c8a64b0c6fda"} Dec 05 08:28:16 crc kubenswrapper[4876]: I1205 08:28:16.198456 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9n88g" event={"ID":"2739f16b-aa7a-45fd-bee3-e69f221176fd","Type":"ContainerStarted","Data":"43b6bf2f6b5932f0e91b8dccb87c25ff11d801136a485a1ea30587f0b3f40fc6"} Dec 05 08:28:16 crc kubenswrapper[4876]: I1205 08:28:16.224221 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9n88g" podStartSLOduration=2.786107202 podStartE2EDuration="4.224197151s" podCreationTimestamp="2025-12-05 08:28:12 +0000 UTC" firstStartedPulling="2025-12-05 08:28:14.183211741 +0000 UTC m=+758.671876413" lastFinishedPulling="2025-12-05 08:28:15.62130172 +0000 UTC m=+760.109966362" observedRunningTime="2025-12-05 08:28:16.218482817 +0000 UTC m=+760.707147459" watchObservedRunningTime="2025-12-05 08:28:16.224197151 +0000 UTC m=+760.712861783" Dec 05 08:28:22 crc kubenswrapper[4876]: I1205 08:28:22.799363 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:22 crc kubenswrapper[4876]: I1205 08:28:22.799861 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:22 crc kubenswrapper[4876]: I1205 08:28:22.853837 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:23 crc kubenswrapper[4876]: I1205 08:28:23.292228 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:23 crc kubenswrapper[4876]: I1205 08:28:23.348952 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9n88g"] Dec 05 08:28:25 crc kubenswrapper[4876]: I1205 08:28:25.247690 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9n88g" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerName="registry-server" containerID="cri-o://43b6bf2f6b5932f0e91b8dccb87c25ff11d801136a485a1ea30587f0b3f40fc6" gracePeriod=2 Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.264244 4876 generic.go:334] "Generic (PLEG): container finished" podID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerID="43b6bf2f6b5932f0e91b8dccb87c25ff11d801136a485a1ea30587f0b3f40fc6" exitCode=0 Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.264295 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9n88g" event={"ID":"2739f16b-aa7a-45fd-bee3-e69f221176fd","Type":"ContainerDied","Data":"43b6bf2f6b5932f0e91b8dccb87c25ff11d801136a485a1ea30587f0b3f40fc6"} Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.674511 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.735743 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w86hg\" (UniqueName: \"kubernetes.io/projected/2739f16b-aa7a-45fd-bee3-e69f221176fd-kube-api-access-w86hg\") pod \"2739f16b-aa7a-45fd-bee3-e69f221176fd\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.735810 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-utilities\") pod \"2739f16b-aa7a-45fd-bee3-e69f221176fd\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.735834 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-catalog-content\") pod \"2739f16b-aa7a-45fd-bee3-e69f221176fd\" (UID: \"2739f16b-aa7a-45fd-bee3-e69f221176fd\") " Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.738694 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-utilities" (OuterVolumeSpecName: "utilities") pod "2739f16b-aa7a-45fd-bee3-e69f221176fd" (UID: "2739f16b-aa7a-45fd-bee3-e69f221176fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.740895 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2739f16b-aa7a-45fd-bee3-e69f221176fd-kube-api-access-w86hg" (OuterVolumeSpecName: "kube-api-access-w86hg") pod "2739f16b-aa7a-45fd-bee3-e69f221176fd" (UID: "2739f16b-aa7a-45fd-bee3-e69f221176fd"). InnerVolumeSpecName "kube-api-access-w86hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.786828 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2739f16b-aa7a-45fd-bee3-e69f221176fd" (UID: "2739f16b-aa7a-45fd-bee3-e69f221176fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.836803 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w86hg\" (UniqueName: \"kubernetes.io/projected/2739f16b-aa7a-45fd-bee3-e69f221176fd-kube-api-access-w86hg\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.836855 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:26 crc kubenswrapper[4876]: I1205 08:28:26.836865 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2739f16b-aa7a-45fd-bee3-e69f221176fd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:27 crc kubenswrapper[4876]: I1205 08:28:27.271981 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9n88g" event={"ID":"2739f16b-aa7a-45fd-bee3-e69f221176fd","Type":"ContainerDied","Data":"aa830717b9e3cf4692f0e32ced5cf926bb07c5a4c81cd7c81d27b163f72df1b3"} Dec 05 08:28:27 crc kubenswrapper[4876]: I1205 08:28:27.272047 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9n88g" Dec 05 08:28:27 crc kubenswrapper[4876]: I1205 08:28:27.272062 4876 scope.go:117] "RemoveContainer" containerID="43b6bf2f6b5932f0e91b8dccb87c25ff11d801136a485a1ea30587f0b3f40fc6" Dec 05 08:28:27 crc kubenswrapper[4876]: I1205 08:28:27.293700 4876 scope.go:117] "RemoveContainer" containerID="f439b42a17362e20dc03bf17e7a154888f5630d689d6f7251228c8a64b0c6fda" Dec 05 08:28:27 crc kubenswrapper[4876]: I1205 08:28:27.303859 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9n88g"] Dec 05 08:28:27 crc kubenswrapper[4876]: I1205 08:28:27.316001 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9n88g"] Dec 05 08:28:27 crc kubenswrapper[4876]: I1205 08:28:27.328369 4876 scope.go:117] "RemoveContainer" containerID="02aed35fcf66b7eb8133f3490b0d94a933aec962ff7c55cfa69619fa6b5a4c0f" Dec 05 08:28:27 crc kubenswrapper[4876]: I1205 08:28:27.828968 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" path="/var/lib/kubelet/pods/2739f16b-aa7a-45fd-bee3-e69f221176fd/volumes" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.287324 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-gwp9v"] Dec 05 08:28:28 crc kubenswrapper[4876]: E1205 08:28:28.287609 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerName="extract-utilities" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.287631 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerName="extract-utilities" Dec 05 08:28:28 crc kubenswrapper[4876]: E1205 08:28:28.287659 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerName="registry-server" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.287670 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerName="registry-server" Dec 05 08:28:28 crc kubenswrapper[4876]: E1205 08:28:28.287696 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerName="extract-content" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.287707 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerName="extract-content" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.287853 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2739f16b-aa7a-45fd-bee3-e69f221176fd" containerName="registry-server" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.288473 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-gwp9v" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.291920 4876 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-zhbfl" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.292250 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.292269 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.298037 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-qrhd7"] Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.298817 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-qrhd7" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.303060 4876 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-bqqz8" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.303668 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-gwp9v"] Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.308062 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-clclm"] Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.308687 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.310954 4876 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jl4t9" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.315356 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-qrhd7"] Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.325719 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-clclm"] Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.360748 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqvdm\" (UniqueName: \"kubernetes.io/projected/398bf935-c5be-48a2-a68f-f884fa8dac57-kube-api-access-zqvdm\") pod \"cert-manager-webhook-5655c58dd6-clclm\" (UID: \"398bf935-c5be-48a2-a68f-f884fa8dac57\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.360797 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj6vk\" (UniqueName: \"kubernetes.io/projected/cdad8ccc-4995-4398-b83e-10a97b7187c9-kube-api-access-wj6vk\") pod \"cert-manager-cainjector-7f985d654d-gwp9v\" (UID: \"cdad8ccc-4995-4398-b83e-10a97b7187c9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-gwp9v" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.360862 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8sgp\" (UniqueName: \"kubernetes.io/projected/ac3650c2-046a-4e3a-a5ce-f0b4b293efae-kube-api-access-d8sgp\") pod \"cert-manager-5b446d88c5-qrhd7\" (UID: \"ac3650c2-046a-4e3a-a5ce-f0b4b293efae\") " pod="cert-manager/cert-manager-5b446d88c5-qrhd7" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.462715 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj6vk\" (UniqueName: \"kubernetes.io/projected/cdad8ccc-4995-4398-b83e-10a97b7187c9-kube-api-access-wj6vk\") pod \"cert-manager-cainjector-7f985d654d-gwp9v\" (UID: \"cdad8ccc-4995-4398-b83e-10a97b7187c9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-gwp9v" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.462768 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqvdm\" (UniqueName: \"kubernetes.io/projected/398bf935-c5be-48a2-a68f-f884fa8dac57-kube-api-access-zqvdm\") pod \"cert-manager-webhook-5655c58dd6-clclm\" (UID: \"398bf935-c5be-48a2-a68f-f884fa8dac57\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.462853 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8sgp\" (UniqueName: \"kubernetes.io/projected/ac3650c2-046a-4e3a-a5ce-f0b4b293efae-kube-api-access-d8sgp\") pod \"cert-manager-5b446d88c5-qrhd7\" (UID: \"ac3650c2-046a-4e3a-a5ce-f0b4b293efae\") " pod="cert-manager/cert-manager-5b446d88c5-qrhd7" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.482776 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj6vk\" (UniqueName: \"kubernetes.io/projected/cdad8ccc-4995-4398-b83e-10a97b7187c9-kube-api-access-wj6vk\") pod \"cert-manager-cainjector-7f985d654d-gwp9v\" (UID: \"cdad8ccc-4995-4398-b83e-10a97b7187c9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-gwp9v" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.484170 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8sgp\" (UniqueName: \"kubernetes.io/projected/ac3650c2-046a-4e3a-a5ce-f0b4b293efae-kube-api-access-d8sgp\") pod \"cert-manager-5b446d88c5-qrhd7\" (UID: \"ac3650c2-046a-4e3a-a5ce-f0b4b293efae\") " pod="cert-manager/cert-manager-5b446d88c5-qrhd7" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.486609 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqvdm\" (UniqueName: \"kubernetes.io/projected/398bf935-c5be-48a2-a68f-f884fa8dac57-kube-api-access-zqvdm\") pod \"cert-manager-webhook-5655c58dd6-clclm\" (UID: \"398bf935-c5be-48a2-a68f-f884fa8dac57\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.604818 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-gwp9v" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.619889 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-qrhd7" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.628991 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.810163 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-clclm"] Dec 05 08:28:28 crc kubenswrapper[4876]: I1205 08:28:28.863679 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-qrhd7"] Dec 05 08:28:28 crc kubenswrapper[4876]: W1205 08:28:28.870963 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac3650c2_046a_4e3a_a5ce_f0b4b293efae.slice/crio-b2c6626d15b4022a18608b31ff41db67835415f5dc556b0ce09f94e6115e5a15 WatchSource:0}: Error finding container b2c6626d15b4022a18608b31ff41db67835415f5dc556b0ce09f94e6115e5a15: Status 404 returned error can't find the container with id b2c6626d15b4022a18608b31ff41db67835415f5dc556b0ce09f94e6115e5a15 Dec 05 08:28:29 crc kubenswrapper[4876]: I1205 08:28:29.086261 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-gwp9v"] Dec 05 08:28:29 crc kubenswrapper[4876]: W1205 08:28:29.090300 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdad8ccc_4995_4398_b83e_10a97b7187c9.slice/crio-670b9394bc57995cd5f08e6bd79b6467b704a2f1f2b710b44b2e0c23205a27c8 WatchSource:0}: Error finding container 670b9394bc57995cd5f08e6bd79b6467b704a2f1f2b710b44b2e0c23205a27c8: Status 404 returned error can't find the container with id 670b9394bc57995cd5f08e6bd79b6467b704a2f1f2b710b44b2e0c23205a27c8 Dec 05 08:28:29 crc kubenswrapper[4876]: I1205 08:28:29.286236 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-gwp9v" event={"ID":"cdad8ccc-4995-4398-b83e-10a97b7187c9","Type":"ContainerStarted","Data":"670b9394bc57995cd5f08e6bd79b6467b704a2f1f2b710b44b2e0c23205a27c8"} Dec 05 08:28:29 crc kubenswrapper[4876]: I1205 08:28:29.287254 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-qrhd7" event={"ID":"ac3650c2-046a-4e3a-a5ce-f0b4b293efae","Type":"ContainerStarted","Data":"b2c6626d15b4022a18608b31ff41db67835415f5dc556b0ce09f94e6115e5a15"} Dec 05 08:28:29 crc kubenswrapper[4876]: I1205 08:28:29.287847 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" event={"ID":"398bf935-c5be-48a2-a68f-f884fa8dac57","Type":"ContainerStarted","Data":"792859be95cd164288af362bb9732fcbc2c56e77ea06e8dcf43b1dd01f55661a"} Dec 05 08:28:32 crc kubenswrapper[4876]: I1205 08:28:32.304525 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" event={"ID":"398bf935-c5be-48a2-a68f-f884fa8dac57","Type":"ContainerStarted","Data":"fa0a85425044ac134829c35de9f903bee06e66ebad3f7c66d4be5ed056e9deda"} Dec 05 08:28:32 crc kubenswrapper[4876]: I1205 08:28:32.305053 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" Dec 05 08:28:32 crc kubenswrapper[4876]: I1205 08:28:32.306984 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-gwp9v" event={"ID":"cdad8ccc-4995-4398-b83e-10a97b7187c9","Type":"ContainerStarted","Data":"04b1677e9f400794da37c75efd0660d5b5919cc80945420fac0b7ee25baa2434"} Dec 05 08:28:32 crc kubenswrapper[4876]: I1205 08:28:32.311951 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-qrhd7" event={"ID":"ac3650c2-046a-4e3a-a5ce-f0b4b293efae","Type":"ContainerStarted","Data":"2ca393b980fa16474942a491a63dee6f0127ca0c7041f300b9cd9c17276f7744"} Dec 05 08:28:32 crc kubenswrapper[4876]: I1205 08:28:32.329031 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" podStartSLOduration=1.325282297 podStartE2EDuration="4.329013704s" podCreationTimestamp="2025-12-05 08:28:28 +0000 UTC" firstStartedPulling="2025-12-05 08:28:28.827798512 +0000 UTC m=+773.316463134" lastFinishedPulling="2025-12-05 08:28:31.831529919 +0000 UTC m=+776.320194541" observedRunningTime="2025-12-05 08:28:32.326242389 +0000 UTC m=+776.814907011" watchObservedRunningTime="2025-12-05 08:28:32.329013704 +0000 UTC m=+776.817678326" Dec 05 08:28:32 crc kubenswrapper[4876]: I1205 08:28:32.342015 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-gwp9v" podStartSLOduration=1.469443381 podStartE2EDuration="4.341996762s" podCreationTimestamp="2025-12-05 08:28:28 +0000 UTC" firstStartedPulling="2025-12-05 08:28:29.092126918 +0000 UTC m=+773.580791530" lastFinishedPulling="2025-12-05 08:28:31.964680289 +0000 UTC m=+776.453344911" observedRunningTime="2025-12-05 08:28:32.339503465 +0000 UTC m=+776.828168087" watchObservedRunningTime="2025-12-05 08:28:32.341996762 +0000 UTC m=+776.830661384" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.119848 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-qrhd7" podStartSLOduration=7.032925084 podStartE2EDuration="10.119823521s" podCreationTimestamp="2025-12-05 08:28:28 +0000 UTC" firstStartedPulling="2025-12-05 08:28:28.872741207 +0000 UTC m=+773.361405829" lastFinishedPulling="2025-12-05 08:28:31.959639624 +0000 UTC m=+776.448304266" observedRunningTime="2025-12-05 08:28:32.358044432 +0000 UTC m=+776.846709054" watchObservedRunningTime="2025-12-05 08:28:38.119823521 +0000 UTC m=+782.608488173" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.124495 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-58npg"] Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.125302 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovn-controller" containerID="cri-o://db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d" gracePeriod=30 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.125990 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="sbdb" containerID="cri-o://375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874" gracePeriod=30 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.126101 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="nbdb" containerID="cri-o://22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9" gracePeriod=30 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.126250 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="northd" containerID="cri-o://47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9" gracePeriod=30 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.126338 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovn-acl-logging" containerID="cri-o://d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415" gracePeriod=30 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.126366 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kube-rbac-proxy-node" containerID="cri-o://7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795" gracePeriod=30 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.126552 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16" gracePeriod=30 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.165810 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" containerID="cri-o://986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9" gracePeriod=30 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.345085 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovnkube-controller/3.log" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.347698 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovn-acl-logging/0.log" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348045 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovn-controller/0.log" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348305 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9" exitCode=0 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348329 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16" exitCode=0 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348338 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795" exitCode=0 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348346 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415" exitCode=143 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348352 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d" exitCode=143 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348383 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9"} Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348407 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16"} Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348416 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795"} Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348425 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415"} Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348434 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d"} Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.348449 4876 scope.go:117] "RemoveContainer" containerID="f60c1f05054fe346feeccd4042e830ff2266c891c042bce461bfe8ac8d1071d1" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.354600 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/2.log" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.355111 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/1.log" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.355149 4876 generic.go:334] "Generic (PLEG): container finished" podID="e7274e0d-7f30-495c-800a-5dcbf1656b0d" containerID="341c70583ca8b1edc982af298aae9fa74350b607faec9525aa3fe1a89c48e098" exitCode=2 Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.355181 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jwrfx" event={"ID":"e7274e0d-7f30-495c-800a-5dcbf1656b0d","Type":"ContainerDied","Data":"341c70583ca8b1edc982af298aae9fa74350b607faec9525aa3fe1a89c48e098"} Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.355695 4876 scope.go:117] "RemoveContainer" containerID="341c70583ca8b1edc982af298aae9fa74350b607faec9525aa3fe1a89c48e098" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.417221 4876 scope.go:117] "RemoveContainer" containerID="bfaef552f975637e73126b9c0d5d692ce06c03fef3cf37929875c4111ece26b3" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.476699 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovn-acl-logging/0.log" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.477303 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovn-controller/0.log" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.477954 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490226 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-log-socket\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490258 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcg2f\" (UniqueName: \"kubernetes.io/projected/446e4833-20af-47c0-b4e0-7116c3972366-kube-api-access-fcg2f\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490289 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-var-lib-cni-networks-ovn-kubernetes\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490308 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/446e4833-20af-47c0-b4e0-7116c3972366-ovn-node-metrics-cert\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490321 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-ovn-kubernetes\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490340 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-config\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490352 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-systemd\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490369 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-bin\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490383 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-kubelet\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490399 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-systemd-units\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490418 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-node-log\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490429 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-netd\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490443 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-openvswitch\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490458 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-etc-openvswitch\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490477 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-slash\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490492 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-netns\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490504 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-var-lib-openvswitch\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490517 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-script-lib\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490538 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-ovn\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.490555 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-env-overrides\") pod \"446e4833-20af-47c0-b4e0-7116c3972366\" (UID: \"446e4833-20af-47c0-b4e0-7116c3972366\") " Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491078 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491109 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-log-socket" (OuterVolumeSpecName: "log-socket") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491336 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-node-log" (OuterVolumeSpecName: "node-log") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491357 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-slash" (OuterVolumeSpecName: "host-slash") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491385 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491406 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491426 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491433 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491481 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491458 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491512 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491520 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491522 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491535 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491543 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.491852 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.492146 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.500031 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/446e4833-20af-47c0-b4e0-7116c3972366-kube-api-access-fcg2f" (OuterVolumeSpecName: "kube-api-access-fcg2f") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "kube-api-access-fcg2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.501996 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446e4833-20af-47c0-b4e0-7116c3972366-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.517116 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "446e4833-20af-47c0-b4e0-7116c3972366" (UID: "446e4833-20af-47c0-b4e0-7116c3972366"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534467 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8gj8z"] Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534697 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534717 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534730 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kube-rbac-proxy-node" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534738 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kube-rbac-proxy-node" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534753 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="northd" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534764 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="northd" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534774 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534782 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534793 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovn-acl-logging" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534801 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovn-acl-logging" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534816 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="nbdb" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534842 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="nbdb" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534855 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kubecfg-setup" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534862 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kubecfg-setup" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534871 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534879 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534893 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534917 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534929 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534937 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534952 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovn-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534960 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovn-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.534970 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="sbdb" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.534977 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="sbdb" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535090 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535104 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535118 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535128 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535138 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="sbdb" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535147 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kube-rbac-proxy-node" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535160 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535169 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovn-acl-logging" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535179 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="northd" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535190 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="nbdb" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535202 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovn-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: E1205 08:28:38.535308 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535319 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.535435 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="446e4833-20af-47c0-b4e0-7116c3972366" containerName="ovnkube-controller" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.537396 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.591865 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-node-log\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.591985 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592017 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-kubelet\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592045 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-ovnkube-config\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592100 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-systemd\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592136 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-ovnkube-script-lib\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592164 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eca8b039-3498-47b1-9c42-03f0fb479317-ovn-node-metrics-cert\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592185 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-var-lib-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592205 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-log-socket\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592225 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-cni-netd\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592257 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-run-netns\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592279 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99pbp\" (UniqueName: \"kubernetes.io/projected/eca8b039-3498-47b1-9c42-03f0fb479317-kube-api-access-99pbp\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592305 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592333 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-cni-bin\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592361 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-slash\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592381 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-env-overrides\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592444 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-etc-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592490 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-ovn\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592505 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-run-ovn-kubernetes\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592531 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-systemd-units\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592611 4876 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592622 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592633 4876 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592642 4876 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592651 4876 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-log-socket\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592659 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcg2f\" (UniqueName: \"kubernetes.io/projected/446e4833-20af-47c0-b4e0-7116c3972366-kube-api-access-fcg2f\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592669 4876 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592678 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/446e4833-20af-47c0-b4e0-7116c3972366-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592687 4876 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592695 4876 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/446e4833-20af-47c0-b4e0-7116c3972366-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592703 4876 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592710 4876 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592718 4876 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592727 4876 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592735 4876 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-node-log\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592743 4876 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592750 4876 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592758 4876 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592767 4876 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-slash\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.592774 4876 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/446e4833-20af-47c0-b4e0-7116c3972366-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.632093 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-clclm" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693469 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-ovnkube-config\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693514 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-systemd\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693561 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-ovnkube-script-lib\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693585 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eca8b039-3498-47b1-9c42-03f0fb479317-ovn-node-metrics-cert\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693608 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-var-lib-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693642 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-log-socket\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693648 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-systemd\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693704 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-cni-netd\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693661 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-cni-netd\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693744 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-log-socket\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693769 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-run-netns\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693750 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-run-netns\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693808 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99pbp\" (UniqueName: \"kubernetes.io/projected/eca8b039-3498-47b1-9c42-03f0fb479317-kube-api-access-99pbp\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693838 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693864 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-cni-bin\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693893 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-slash\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693937 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-etc-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693960 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-env-overrides\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693983 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-ovn\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694003 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-run-ovn-kubernetes\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694045 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-systemd-units\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694077 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-node-log\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694103 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.693713 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-var-lib-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694140 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-kubelet\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694161 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-ovn\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694193 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-slash\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694199 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-cni-bin\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694217 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-etc-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694367 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-ovnkube-script-lib\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694403 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-run-ovn-kubernetes\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694435 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694431 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-ovnkube-config\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694500 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-systemd-units\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694532 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-host-kubelet\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694547 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eca8b039-3498-47b1-9c42-03f0fb479317-env-overrides\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694555 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-node-log\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.694576 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/eca8b039-3498-47b1-9c42-03f0fb479317-run-openvswitch\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.697316 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eca8b039-3498-47b1-9c42-03f0fb479317-ovn-node-metrics-cert\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.708559 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99pbp\" (UniqueName: \"kubernetes.io/projected/eca8b039-3498-47b1-9c42-03f0fb479317-kube-api-access-99pbp\") pod \"ovnkube-node-8gj8z\" (UID: \"eca8b039-3498-47b1-9c42-03f0fb479317\") " pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: I1205 08:28:38.850247 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:38 crc kubenswrapper[4876]: W1205 08:28:38.880683 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeca8b039_3498_47b1_9c42_03f0fb479317.slice/crio-201bbd6a7d1d58be938325b1b65cc25890c0f708f061ed930b260321bf42c8b2 WatchSource:0}: Error finding container 201bbd6a7d1d58be938325b1b65cc25890c0f708f061ed930b260321bf42c8b2: Status 404 returned error can't find the container with id 201bbd6a7d1d58be938325b1b65cc25890c0f708f061ed930b260321bf42c8b2 Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.362058 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jwrfx_e7274e0d-7f30-495c-800a-5dcbf1656b0d/kube-multus/2.log" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.362193 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jwrfx" event={"ID":"e7274e0d-7f30-495c-800a-5dcbf1656b0d","Type":"ContainerStarted","Data":"b4082d9f99fb4f00bcdc1ecd38b50eaec008679545efc51ea4f3be7d10749449"} Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.364117 4876 generic.go:334] "Generic (PLEG): container finished" podID="eca8b039-3498-47b1-9c42-03f0fb479317" containerID="34cf65106f1f318f526a8b17c946c1d67a19c52dd2bed290fc3a979478e27537" exitCode=0 Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.364208 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerDied","Data":"34cf65106f1f318f526a8b17c946c1d67a19c52dd2bed290fc3a979478e27537"} Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.364384 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"201bbd6a7d1d58be938325b1b65cc25890c0f708f061ed930b260321bf42c8b2"} Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.369201 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovn-acl-logging/0.log" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.371939 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-58npg_446e4833-20af-47c0-b4e0-7116c3972366/ovn-controller/0.log" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372374 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874" exitCode=0 Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372413 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9" exitCode=0 Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372429 4876 generic.go:334] "Generic (PLEG): container finished" podID="446e4833-20af-47c0-b4e0-7116c3972366" containerID="47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9" exitCode=0 Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372466 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874"} Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372490 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372507 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9"} Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372528 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9"} Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372548 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-58npg" event={"ID":"446e4833-20af-47c0-b4e0-7116c3972366","Type":"ContainerDied","Data":"2a28c1f10fd27337083b72ff07e2ed349bceaf0d312551f0fa4d8a6d7c2fc807"} Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.372598 4876 scope.go:117] "RemoveContainer" containerID="986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.397692 4876 scope.go:117] "RemoveContainer" containerID="375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.436003 4876 scope.go:117] "RemoveContainer" containerID="22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.447592 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-58npg"] Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.450688 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-58npg"] Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.453725 4876 scope.go:117] "RemoveContainer" containerID="47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.478414 4876 scope.go:117] "RemoveContainer" containerID="e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.499934 4876 scope.go:117] "RemoveContainer" containerID="7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.511646 4876 scope.go:117] "RemoveContainer" containerID="d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.529130 4876 scope.go:117] "RemoveContainer" containerID="db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.545138 4876 scope.go:117] "RemoveContainer" containerID="3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.575641 4876 scope.go:117] "RemoveContainer" containerID="986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.576262 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9\": container with ID starting with 986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9 not found: ID does not exist" containerID="986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.576302 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9"} err="failed to get container status \"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9\": rpc error: code = NotFound desc = could not find container \"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9\": container with ID starting with 986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.576323 4876 scope.go:117] "RemoveContainer" containerID="375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.576727 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\": container with ID starting with 375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874 not found: ID does not exist" containerID="375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.576756 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874"} err="failed to get container status \"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\": rpc error: code = NotFound desc = could not find container \"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\": container with ID starting with 375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.576780 4876 scope.go:117] "RemoveContainer" containerID="22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.577161 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\": container with ID starting with 22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9 not found: ID does not exist" containerID="22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.577186 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9"} err="failed to get container status \"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\": rpc error: code = NotFound desc = could not find container \"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\": container with ID starting with 22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.577210 4876 scope.go:117] "RemoveContainer" containerID="47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.577465 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\": container with ID starting with 47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9 not found: ID does not exist" containerID="47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.577506 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9"} err="failed to get container status \"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\": rpc error: code = NotFound desc = could not find container \"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\": container with ID starting with 47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.577529 4876 scope.go:117] "RemoveContainer" containerID="e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.577746 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\": container with ID starting with e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16 not found: ID does not exist" containerID="e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.577766 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16"} err="failed to get container status \"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\": rpc error: code = NotFound desc = could not find container \"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\": container with ID starting with e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.577778 4876 scope.go:117] "RemoveContainer" containerID="7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.578079 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\": container with ID starting with 7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795 not found: ID does not exist" containerID="7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.578134 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795"} err="failed to get container status \"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\": rpc error: code = NotFound desc = could not find container \"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\": container with ID starting with 7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.578171 4876 scope.go:117] "RemoveContainer" containerID="d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.578502 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\": container with ID starting with d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415 not found: ID does not exist" containerID="d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.578526 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415"} err="failed to get container status \"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\": rpc error: code = NotFound desc = could not find container \"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\": container with ID starting with d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.578540 4876 scope.go:117] "RemoveContainer" containerID="db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.578788 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\": container with ID starting with db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d not found: ID does not exist" containerID="db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.578811 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d"} err="failed to get container status \"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\": rpc error: code = NotFound desc = could not find container \"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\": container with ID starting with db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.578826 4876 scope.go:117] "RemoveContainer" containerID="3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5" Dec 05 08:28:39 crc kubenswrapper[4876]: E1205 08:28:39.579075 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\": container with ID starting with 3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5 not found: ID does not exist" containerID="3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.579105 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5"} err="failed to get container status \"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\": rpc error: code = NotFound desc = could not find container \"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\": container with ID starting with 3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.579124 4876 scope.go:117] "RemoveContainer" containerID="986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.579329 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9"} err="failed to get container status \"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9\": rpc error: code = NotFound desc = could not find container \"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9\": container with ID starting with 986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.579347 4876 scope.go:117] "RemoveContainer" containerID="375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.579545 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874"} err="failed to get container status \"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\": rpc error: code = NotFound desc = could not find container \"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\": container with ID starting with 375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.579561 4876 scope.go:117] "RemoveContainer" containerID="22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.579724 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9"} err="failed to get container status \"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\": rpc error: code = NotFound desc = could not find container \"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\": container with ID starting with 22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.579741 4876 scope.go:117] "RemoveContainer" containerID="47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.580141 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9"} err="failed to get container status \"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\": rpc error: code = NotFound desc = could not find container \"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\": container with ID starting with 47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.580161 4876 scope.go:117] "RemoveContainer" containerID="e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.580407 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16"} err="failed to get container status \"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\": rpc error: code = NotFound desc = could not find container \"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\": container with ID starting with e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.580427 4876 scope.go:117] "RemoveContainer" containerID="7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.580638 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795"} err="failed to get container status \"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\": rpc error: code = NotFound desc = could not find container \"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\": container with ID starting with 7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.580655 4876 scope.go:117] "RemoveContainer" containerID="d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.580873 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415"} err="failed to get container status \"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\": rpc error: code = NotFound desc = could not find container \"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\": container with ID starting with d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.580913 4876 scope.go:117] "RemoveContainer" containerID="db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.581130 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d"} err="failed to get container status \"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\": rpc error: code = NotFound desc = could not find container \"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\": container with ID starting with db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.581156 4876 scope.go:117] "RemoveContainer" containerID="3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.581521 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5"} err="failed to get container status \"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\": rpc error: code = NotFound desc = could not find container \"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\": container with ID starting with 3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.581539 4876 scope.go:117] "RemoveContainer" containerID="986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.581776 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9"} err="failed to get container status \"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9\": rpc error: code = NotFound desc = could not find container \"986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9\": container with ID starting with 986466eeb5fc5553ddfe0f6d5bbed15c7e7e579c692c8871f738a0a850c876a9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.581796 4876 scope.go:117] "RemoveContainer" containerID="375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.582020 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874"} err="failed to get container status \"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\": rpc error: code = NotFound desc = could not find container \"375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874\": container with ID starting with 375977f6b74858165040b1196ee253ef136bdb0b3298086256fa4d5f4315d874 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.582038 4876 scope.go:117] "RemoveContainer" containerID="22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.582282 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9"} err="failed to get container status \"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\": rpc error: code = NotFound desc = could not find container \"22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9\": container with ID starting with 22269426a2718d3cf681edc5cc7e9116d16276be5120efa1c0d7e91033bafca9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.582301 4876 scope.go:117] "RemoveContainer" containerID="47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.582516 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9"} err="failed to get container status \"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\": rpc error: code = NotFound desc = could not find container \"47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9\": container with ID starting with 47a494f4a385607ba7eb83fbae2daad09f4d89efe409ba0b328d5120c5bdbcf9 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.582544 4876 scope.go:117] "RemoveContainer" containerID="e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.582773 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16"} err="failed to get container status \"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\": rpc error: code = NotFound desc = could not find container \"e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16\": container with ID starting with e14ac408dd4a5c6b128f62b6dc629c7222ec92e520575a4c599d9d7687e3ad16 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.582791 4876 scope.go:117] "RemoveContainer" containerID="7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.583020 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795"} err="failed to get container status \"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\": rpc error: code = NotFound desc = could not find container \"7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795\": container with ID starting with 7ebc5086bde9138ef0358820d30f222f9ca010e803b9a9d79c39eeb0e078b795 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.583053 4876 scope.go:117] "RemoveContainer" containerID="d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.583291 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415"} err="failed to get container status \"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\": rpc error: code = NotFound desc = could not find container \"d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415\": container with ID starting with d9625e79f68e4458b0b56c15e60cbbf36cc84603c839fc467281414095e90415 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.583312 4876 scope.go:117] "RemoveContainer" containerID="db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.583526 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d"} err="failed to get container status \"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\": rpc error: code = NotFound desc = could not find container \"db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d\": container with ID starting with db5c0e2fe6e10233fb3679bfd20c4334d2905351b85be4ffebf1266e2d96f79d not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.583552 4876 scope.go:117] "RemoveContainer" containerID="3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.583791 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5"} err="failed to get container status \"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\": rpc error: code = NotFound desc = could not find container \"3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5\": container with ID starting with 3ffcd4508f9873e79959f177b509a70801d07c318d6c04350c7fbbbfb4046da5 not found: ID does not exist" Dec 05 08:28:39 crc kubenswrapper[4876]: I1205 08:28:39.835893 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="446e4833-20af-47c0-b4e0-7116c3972366" path="/var/lib/kubelet/pods/446e4833-20af-47c0-b4e0-7116c3972366/volumes" Dec 05 08:28:40 crc kubenswrapper[4876]: I1205 08:28:40.385209 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"76216af5a94fc4c755391d0d8bcb231c2d009dfd38ecbf9e9944ae61a14a02c1"} Dec 05 08:28:40 crc kubenswrapper[4876]: I1205 08:28:40.385271 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"87a83673e260bac4857df5cae3eae4caab6237fc9c3b36f9eca6bb9cea8f50bc"} Dec 05 08:28:40 crc kubenswrapper[4876]: I1205 08:28:40.385290 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"356d09a4ee8f786695cf3e1d0ab7c9853a2aaa65246769aa15cf635fcb56aae9"} Dec 05 08:28:40 crc kubenswrapper[4876]: I1205 08:28:40.385308 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"58c4edc826b4587d019839725dc857519d3bb89d81951ec499127e42a231b350"} Dec 05 08:28:40 crc kubenswrapper[4876]: I1205 08:28:40.385324 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"f112d4ededa02f2bad1e73c7ffdf96bbd76e23c165bb04957e037558bf9d0ac9"} Dec 05 08:28:40 crc kubenswrapper[4876]: I1205 08:28:40.385342 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"aa326f8e3523aa4694ca7f906a9b9fa1fd2fdb6d349bc047563f6b4aa2a62446"} Dec 05 08:28:42 crc kubenswrapper[4876]: I1205 08:28:42.403734 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"2458cc46bae2fa753452561e5c8d2bb890354ecf9db067b85461131a5cfa7ea2"} Dec 05 08:28:45 crc kubenswrapper[4876]: I1205 08:28:45.422565 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" event={"ID":"eca8b039-3498-47b1-9c42-03f0fb479317","Type":"ContainerStarted","Data":"79f8b2308cab8f2075e042a112d5e07b5fd4b0bb8208bea4ec039ac201883fe7"} Dec 05 08:28:46 crc kubenswrapper[4876]: I1205 08:28:46.427304 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:46 crc kubenswrapper[4876]: I1205 08:28:46.457205 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" podStartSLOduration=8.457184927 podStartE2EDuration="8.457184927s" podCreationTimestamp="2025-12-05 08:28:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:28:46.452678176 +0000 UTC m=+790.941342808" watchObservedRunningTime="2025-12-05 08:28:46.457184927 +0000 UTC m=+790.945849569" Dec 05 08:28:46 crc kubenswrapper[4876]: I1205 08:28:46.461213 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:47 crc kubenswrapper[4876]: I1205 08:28:47.432927 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:47 crc kubenswrapper[4876]: I1205 08:28:47.433279 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:47 crc kubenswrapper[4876]: I1205 08:28:47.467184 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:28:49 crc kubenswrapper[4876]: I1205 08:28:49.465632 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8gj8z" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.203984 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt"] Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.206073 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.208552 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.230640 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt"] Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.271864 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.272140 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.272280 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2smsb\" (UniqueName: \"kubernetes.io/projected/42f3432a-0028-45dc-a0d3-a886bd2da16a-kube-api-access-2smsb\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.373799 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.373990 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2smsb\" (UniqueName: \"kubernetes.io/projected/42f3432a-0028-45dc-a0d3-a886bd2da16a-kube-api-access-2smsb\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.374102 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.375185 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.375404 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.395867 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2smsb\" (UniqueName: \"kubernetes.io/projected/42f3432a-0028-45dc-a0d3-a886bd2da16a-kube-api-access-2smsb\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:17 crc kubenswrapper[4876]: I1205 08:29:17.526043 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:18 crc kubenswrapper[4876]: I1205 08:29:18.109967 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt"] Dec 05 08:29:18 crc kubenswrapper[4876]: W1205 08:29:18.115759 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42f3432a_0028_45dc_a0d3_a886bd2da16a.slice/crio-a0242735543645c59b8c2fdf553231070c3a6faf8a2bc3cea11332e1ee38ed84 WatchSource:0}: Error finding container a0242735543645c59b8c2fdf553231070c3a6faf8a2bc3cea11332e1ee38ed84: Status 404 returned error can't find the container with id a0242735543645c59b8c2fdf553231070c3a6faf8a2bc3cea11332e1ee38ed84 Dec 05 08:29:18 crc kubenswrapper[4876]: I1205 08:29:18.607130 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" event={"ID":"42f3432a-0028-45dc-a0d3-a886bd2da16a","Type":"ContainerStarted","Data":"5c4deebc4490b3c0d4fc78e250c56638f73756c9ccdf32c8f74122d92103e66e"} Dec 05 08:29:18 crc kubenswrapper[4876]: I1205 08:29:18.607982 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" event={"ID":"42f3432a-0028-45dc-a0d3-a886bd2da16a","Type":"ContainerStarted","Data":"a0242735543645c59b8c2fdf553231070c3a6faf8a2bc3cea11332e1ee38ed84"} Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.514858 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-plkcw"] Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.517429 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.525843 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-plkcw"] Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.613192 4876 generic.go:334] "Generic (PLEG): container finished" podID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerID="5c4deebc4490b3c0d4fc78e250c56638f73756c9ccdf32c8f74122d92103e66e" exitCode=0 Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.613244 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" event={"ID":"42f3432a-0028-45dc-a0d3-a886bd2da16a","Type":"ContainerDied","Data":"5c4deebc4490b3c0d4fc78e250c56638f73756c9ccdf32c8f74122d92103e66e"} Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.708634 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-utilities\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.708744 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-catalog-content\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.708803 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpccz\" (UniqueName: \"kubernetes.io/projected/96310d2c-83ba-4695-afaa-6a9e2490b066-kube-api-access-bpccz\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.809305 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-utilities\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.809359 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-catalog-content\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.809387 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpccz\" (UniqueName: \"kubernetes.io/projected/96310d2c-83ba-4695-afaa-6a9e2490b066-kube-api-access-bpccz\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.809964 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-utilities\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.810018 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-catalog-content\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.835890 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpccz\" (UniqueName: \"kubernetes.io/projected/96310d2c-83ba-4695-afaa-6a9e2490b066-kube-api-access-bpccz\") pod \"redhat-operators-plkcw\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:19 crc kubenswrapper[4876]: I1205 08:29:19.852550 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:20 crc kubenswrapper[4876]: I1205 08:29:20.322981 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-plkcw"] Dec 05 08:29:20 crc kubenswrapper[4876]: W1205 08:29:20.329137 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96310d2c_83ba_4695_afaa_6a9e2490b066.slice/crio-09d7e259b51fd7ea1595bf8df0306a033c81935a24375868eb28ab3962c220a6 WatchSource:0}: Error finding container 09d7e259b51fd7ea1595bf8df0306a033c81935a24375868eb28ab3962c220a6: Status 404 returned error can't find the container with id 09d7e259b51fd7ea1595bf8df0306a033c81935a24375868eb28ab3962c220a6 Dec 05 08:29:20 crc kubenswrapper[4876]: I1205 08:29:20.619012 4876 generic.go:334] "Generic (PLEG): container finished" podID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerID="5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750" exitCode=0 Dec 05 08:29:20 crc kubenswrapper[4876]: I1205 08:29:20.619050 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plkcw" event={"ID":"96310d2c-83ba-4695-afaa-6a9e2490b066","Type":"ContainerDied","Data":"5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750"} Dec 05 08:29:20 crc kubenswrapper[4876]: I1205 08:29:20.619074 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plkcw" event={"ID":"96310d2c-83ba-4695-afaa-6a9e2490b066","Type":"ContainerStarted","Data":"09d7e259b51fd7ea1595bf8df0306a033c81935a24375868eb28ab3962c220a6"} Dec 05 08:29:21 crc kubenswrapper[4876]: I1205 08:29:21.627285 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plkcw" event={"ID":"96310d2c-83ba-4695-afaa-6a9e2490b066","Type":"ContainerStarted","Data":"07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd"} Dec 05 08:29:22 crc kubenswrapper[4876]: I1205 08:29:22.637626 4876 generic.go:334] "Generic (PLEG): container finished" podID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerID="5ca4212f8fd4f5893b6b6e1d14ff8e534a3331860de19b15ac0a8fd35ba23ad1" exitCode=0 Dec 05 08:29:22 crc kubenswrapper[4876]: I1205 08:29:22.638806 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" event={"ID":"42f3432a-0028-45dc-a0d3-a886bd2da16a","Type":"ContainerDied","Data":"5ca4212f8fd4f5893b6b6e1d14ff8e534a3331860de19b15ac0a8fd35ba23ad1"} Dec 05 08:29:23 crc kubenswrapper[4876]: I1205 08:29:23.645074 4876 generic.go:334] "Generic (PLEG): container finished" podID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerID="07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd" exitCode=0 Dec 05 08:29:23 crc kubenswrapper[4876]: I1205 08:29:23.645152 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plkcw" event={"ID":"96310d2c-83ba-4695-afaa-6a9e2490b066","Type":"ContainerDied","Data":"07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd"} Dec 05 08:29:23 crc kubenswrapper[4876]: I1205 08:29:23.650325 4876 generic.go:334] "Generic (PLEG): container finished" podID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerID="d1a781caf2d00dca13a26dd6f1363e73857c74a21424f6c09db20ffbbf8e0eac" exitCode=0 Dec 05 08:29:23 crc kubenswrapper[4876]: I1205 08:29:23.650378 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" event={"ID":"42f3432a-0028-45dc-a0d3-a886bd2da16a","Type":"ContainerDied","Data":"d1a781caf2d00dca13a26dd6f1363e73857c74a21424f6c09db20ffbbf8e0eac"} Dec 05 08:29:24 crc kubenswrapper[4876]: I1205 08:29:24.662244 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plkcw" event={"ID":"96310d2c-83ba-4695-afaa-6a9e2490b066","Type":"ContainerStarted","Data":"ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79"} Dec 05 08:29:24 crc kubenswrapper[4876]: I1205 08:29:24.953445 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:24 crc kubenswrapper[4876]: I1205 08:29:24.972495 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-plkcw" podStartSLOduration=2.251866843 podStartE2EDuration="5.972472551s" podCreationTimestamp="2025-12-05 08:29:19 +0000 UTC" firstStartedPulling="2025-12-05 08:29:20.620127409 +0000 UTC m=+825.108792031" lastFinishedPulling="2025-12-05 08:29:24.340733077 +0000 UTC m=+828.829397739" observedRunningTime="2025-12-05 08:29:24.691649074 +0000 UTC m=+829.180313736" watchObservedRunningTime="2025-12-05 08:29:24.972472551 +0000 UTC m=+829.461137183" Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.005403 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-util\") pod \"42f3432a-0028-45dc-a0d3-a886bd2da16a\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.005631 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2smsb\" (UniqueName: \"kubernetes.io/projected/42f3432a-0028-45dc-a0d3-a886bd2da16a-kube-api-access-2smsb\") pod \"42f3432a-0028-45dc-a0d3-a886bd2da16a\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.005734 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-bundle\") pod \"42f3432a-0028-45dc-a0d3-a886bd2da16a\" (UID: \"42f3432a-0028-45dc-a0d3-a886bd2da16a\") " Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.006435 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-bundle" (OuterVolumeSpecName: "bundle") pod "42f3432a-0028-45dc-a0d3-a886bd2da16a" (UID: "42f3432a-0028-45dc-a0d3-a886bd2da16a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.013196 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f3432a-0028-45dc-a0d3-a886bd2da16a-kube-api-access-2smsb" (OuterVolumeSpecName: "kube-api-access-2smsb") pod "42f3432a-0028-45dc-a0d3-a886bd2da16a" (UID: "42f3432a-0028-45dc-a0d3-a886bd2da16a"). InnerVolumeSpecName "kube-api-access-2smsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.031078 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-util" (OuterVolumeSpecName: "util") pod "42f3432a-0028-45dc-a0d3-a886bd2da16a" (UID: "42f3432a-0028-45dc-a0d3-a886bd2da16a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.107598 4876 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-util\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.107658 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2smsb\" (UniqueName: \"kubernetes.io/projected/42f3432a-0028-45dc-a0d3-a886bd2da16a-kube-api-access-2smsb\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.107681 4876 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42f3432a-0028-45dc-a0d3-a886bd2da16a-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.670932 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" event={"ID":"42f3432a-0028-45dc-a0d3-a886bd2da16a","Type":"ContainerDied","Data":"a0242735543645c59b8c2fdf553231070c3a6faf8a2bc3cea11332e1ee38ed84"} Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.670978 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0242735543645c59b8c2fdf553231070c3a6faf8a2bc3cea11332e1ee38ed84" Dec 05 08:29:25 crc kubenswrapper[4876]: I1205 08:29:25.671122 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.399476 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr"] Dec 05 08:29:27 crc kubenswrapper[4876]: E1205 08:29:27.400000 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerName="util" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.400012 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerName="util" Dec 05 08:29:27 crc kubenswrapper[4876]: E1205 08:29:27.400024 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerName="pull" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.400032 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerName="pull" Dec 05 08:29:27 crc kubenswrapper[4876]: E1205 08:29:27.400044 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerName="extract" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.400052 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerName="extract" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.400162 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f3432a-0028-45dc-a0d3-a886bd2da16a" containerName="extract" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.400611 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.405282 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.405291 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.405531 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-24zv9" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.419721 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr"] Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.435318 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn5sq\" (UniqueName: \"kubernetes.io/projected/9ef2d625-98dd-4299-bee0-848f9571b8c7-kube-api-access-pn5sq\") pod \"nmstate-operator-5b5b58f5c8-p68nr\" (UID: \"9ef2d625-98dd-4299-bee0-848f9571b8c7\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.536650 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn5sq\" (UniqueName: \"kubernetes.io/projected/9ef2d625-98dd-4299-bee0-848f9571b8c7-kube-api-access-pn5sq\") pod \"nmstate-operator-5b5b58f5c8-p68nr\" (UID: \"9ef2d625-98dd-4299-bee0-848f9571b8c7\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.556611 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn5sq\" (UniqueName: \"kubernetes.io/projected/9ef2d625-98dd-4299-bee0-848f9571b8c7-kube-api-access-pn5sq\") pod \"nmstate-operator-5b5b58f5c8-p68nr\" (UID: \"9ef2d625-98dd-4299-bee0-848f9571b8c7\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr" Dec 05 08:29:27 crc kubenswrapper[4876]: I1205 08:29:27.716829 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr" Dec 05 08:29:28 crc kubenswrapper[4876]: I1205 08:29:28.177076 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr"] Dec 05 08:29:28 crc kubenswrapper[4876]: I1205 08:29:28.686316 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr" event={"ID":"9ef2d625-98dd-4299-bee0-848f9571b8c7","Type":"ContainerStarted","Data":"ed6af05c559ab5c3ec0dff1d01ef9a880e9b08bff8c3349f578f218d4489d976"} Dec 05 08:29:29 crc kubenswrapper[4876]: I1205 08:29:29.853100 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:29 crc kubenswrapper[4876]: I1205 08:29:29.853433 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:30 crc kubenswrapper[4876]: I1205 08:29:30.889467 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-plkcw" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="registry-server" probeResult="failure" output=< Dec 05 08:29:30 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 08:29:30 crc kubenswrapper[4876]: > Dec 05 08:29:31 crc kubenswrapper[4876]: I1205 08:29:31.716175 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr" event={"ID":"9ef2d625-98dd-4299-bee0-848f9571b8c7","Type":"ContainerStarted","Data":"a44bc3b072d092a5a91ab08b2cb31f4b8fb83da119bb9acfb76e1112117d02d9"} Dec 05 08:29:31 crc kubenswrapper[4876]: I1205 08:29:31.744590 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-p68nr" podStartSLOduration=1.5498585249999999 podStartE2EDuration="4.744572865s" podCreationTimestamp="2025-12-05 08:29:27 +0000 UTC" firstStartedPulling="2025-12-05 08:29:28.196861742 +0000 UTC m=+832.685526364" lastFinishedPulling="2025-12-05 08:29:31.391576082 +0000 UTC m=+835.880240704" observedRunningTime="2025-12-05 08:29:31.738223475 +0000 UTC m=+836.226888097" watchObservedRunningTime="2025-12-05 08:29:31.744572865 +0000 UTC m=+836.233237487" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.348103 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.349623 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.351702 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-vznj7" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.358622 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.359501 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.361921 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.365494 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.377893 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lnnm\" (UniqueName: \"kubernetes.io/projected/f256549a-a4ac-40d4-b0b9-74e5b888d9f4-kube-api-access-9lnnm\") pod \"nmstate-webhook-5f6d4c5ccb-dhgxp\" (UID: \"f256549a-a4ac-40d4-b0b9-74e5b888d9f4\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.377948 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f256549a-a4ac-40d4-b0b9-74e5b888d9f4-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-dhgxp\" (UID: \"f256549a-a4ac-40d4-b0b9-74e5b888d9f4\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.377976 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5dfh\" (UniqueName: \"kubernetes.io/projected/2fb0c525-e24c-4093-b8f7-8a9e3f788c29-kube-api-access-n5dfh\") pod \"nmstate-metrics-7f946cbc9-n4rqm\" (UID: \"2fb0c525-e24c-4093-b8f7-8a9e3f788c29\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.383969 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-n2snz"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.384601 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.388792 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.477780 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.479109 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lnnm\" (UniqueName: \"kubernetes.io/projected/f256549a-a4ac-40d4-b0b9-74e5b888d9f4-kube-api-access-9lnnm\") pod \"nmstate-webhook-5f6d4c5ccb-dhgxp\" (UID: \"f256549a-a4ac-40d4-b0b9-74e5b888d9f4\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.479159 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f256549a-a4ac-40d4-b0b9-74e5b888d9f4-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-dhgxp\" (UID: \"f256549a-a4ac-40d4-b0b9-74e5b888d9f4\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.479196 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5dfh\" (UniqueName: \"kubernetes.io/projected/2fb0c525-e24c-4093-b8f7-8a9e3f788c29-kube-api-access-n5dfh\") pod \"nmstate-metrics-7f946cbc9-n4rqm\" (UID: \"2fb0c525-e24c-4093-b8f7-8a9e3f788c29\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.479602 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.483639 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-f8bbm" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.483979 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.484141 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.488034 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/f256549a-a4ac-40d4-b0b9-74e5b888d9f4-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-dhgxp\" (UID: \"f256549a-a4ac-40d4-b0b9-74e5b888d9f4\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.490039 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.506406 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5dfh\" (UniqueName: \"kubernetes.io/projected/2fb0c525-e24c-4093-b8f7-8a9e3f788c29-kube-api-access-n5dfh\") pod \"nmstate-metrics-7f946cbc9-n4rqm\" (UID: \"2fb0c525-e24c-4093-b8f7-8a9e3f788c29\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.520167 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lnnm\" (UniqueName: \"kubernetes.io/projected/f256549a-a4ac-40d4-b0b9-74e5b888d9f4-kube-api-access-9lnnm\") pod \"nmstate-webhook-5f6d4c5ccb-dhgxp\" (UID: \"f256549a-a4ac-40d4-b0b9-74e5b888d9f4\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.580817 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/85c45836-2564-48e9-8c2f-33db89131c1b-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.580861 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bntw\" (UniqueName: \"kubernetes.io/projected/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-kube-api-access-4bntw\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.580915 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-nmstate-lock\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.580948 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-dbus-socket\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.580992 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvmqk\" (UniqueName: \"kubernetes.io/projected/85c45836-2564-48e9-8c2f-33db89131c1b-kube-api-access-cvmqk\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.581022 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/85c45836-2564-48e9-8c2f-33db89131c1b-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.581100 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-ovs-socket\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.659732 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-576575b889-kmlg4"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.661852 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.673418 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.676796 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-576575b889-kmlg4"] Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.682078 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/85c45836-2564-48e9-8c2f-33db89131c1b-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.682165 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-ovs-socket\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.682211 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/85c45836-2564-48e9-8c2f-33db89131c1b-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.682232 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bntw\" (UniqueName: \"kubernetes.io/projected/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-kube-api-access-4bntw\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.682255 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-nmstate-lock\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.682287 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-dbus-socket\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.682305 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvmqk\" (UniqueName: \"kubernetes.io/projected/85c45836-2564-48e9-8c2f-33db89131c1b-kube-api-access-cvmqk\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.682785 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-nmstate-lock\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.683071 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-dbus-socket\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: E1205 08:29:37.683140 4876 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 05 08:29:37 crc kubenswrapper[4876]: E1205 08:29:37.683187 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/85c45836-2564-48e9-8c2f-33db89131c1b-plugin-serving-cert podName:85c45836-2564-48e9-8c2f-33db89131c1b nodeName:}" failed. No retries permitted until 2025-12-05 08:29:38.183169718 +0000 UTC m=+842.671834340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/85c45836-2564-48e9-8c2f-33db89131c1b-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-g4pnh" (UID: "85c45836-2564-48e9-8c2f-33db89131c1b") : secret "plugin-serving-cert" not found Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.683378 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-ovs-socket\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.687450 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/85c45836-2564-48e9-8c2f-33db89131c1b-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.687774 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.712271 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvmqk\" (UniqueName: \"kubernetes.io/projected/85c45836-2564-48e9-8c2f-33db89131c1b-kube-api-access-cvmqk\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.713594 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bntw\" (UniqueName: \"kubernetes.io/projected/8c0975c5-f6cd-4565-a5fc-c9247bc8871c-kube-api-access-4bntw\") pod \"nmstate-handler-n2snz\" (UID: \"8c0975c5-f6cd-4565-a5fc-c9247bc8871c\") " pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.791345 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-oauth-config\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.791621 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-config\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.791678 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-service-ca\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.791709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-serving-cert\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.791739 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wqgx\" (UniqueName: \"kubernetes.io/projected/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-kube-api-access-9wqgx\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.791776 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-trusted-ca-bundle\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.791796 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-oauth-serving-cert\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.892532 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-service-ca\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.892584 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-serving-cert\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.892614 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wqgx\" (UniqueName: \"kubernetes.io/projected/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-kube-api-access-9wqgx\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.892706 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-trusted-ca-bundle\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.893735 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-service-ca\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.893851 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-oauth-serving-cert\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.893960 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-oauth-config\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.894147 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-config\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.894174 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-trusted-ca-bundle\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.894838 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-oauth-serving-cert\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.895336 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-config\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.898833 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-oauth-config\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.901762 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-console-serving-cert\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.909041 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wqgx\" (UniqueName: \"kubernetes.io/projected/658d8a8a-3c7c-4e64-a98c-2f619b52c21e-kube-api-access-9wqgx\") pod \"console-576575b889-kmlg4\" (UID: \"658d8a8a-3c7c-4e64-a98c-2f619b52c21e\") " pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:37 crc kubenswrapper[4876]: I1205 08:29:37.974632 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.004974 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:38 crc kubenswrapper[4876]: W1205 08:29:38.024417 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c0975c5_f6cd_4565_a5fc_c9247bc8871c.slice/crio-ee62cd1259e7adc3faca2f43459ce9643e90494dd47065c623993c9221a021ba WatchSource:0}: Error finding container ee62cd1259e7adc3faca2f43459ce9643e90494dd47065c623993c9221a021ba: Status 404 returned error can't find the container with id ee62cd1259e7adc3faca2f43459ce9643e90494dd47065c623993c9221a021ba Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.160974 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm"] Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.183468 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp"] Dec 05 08:29:38 crc kubenswrapper[4876]: W1205 08:29:38.193137 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf256549a_a4ac_40d4_b0b9_74e5b888d9f4.slice/crio-f38788a253f9fc5aecc6ab74c0102b220bda79687c865d45d30d9b7a3f95710c WatchSource:0}: Error finding container f38788a253f9fc5aecc6ab74c0102b220bda79687c865d45d30d9b7a3f95710c: Status 404 returned error can't find the container with id f38788a253f9fc5aecc6ab74c0102b220bda79687c865d45d30d9b7a3f95710c Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.199413 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/85c45836-2564-48e9-8c2f-33db89131c1b-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.204348 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/85c45836-2564-48e9-8c2f-33db89131c1b-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g4pnh\" (UID: \"85c45836-2564-48e9-8c2f-33db89131c1b\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.248618 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-576575b889-kmlg4"] Dec 05 08:29:38 crc kubenswrapper[4876]: W1205 08:29:38.254191 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod658d8a8a_3c7c_4e64_a98c_2f619b52c21e.slice/crio-88a86f87627571f8098ea758f8461535421de9cb4d41b69c490d81caa6caa3ea WatchSource:0}: Error finding container 88a86f87627571f8098ea758f8461535421de9cb4d41b69c490d81caa6caa3ea: Status 404 returned error can't find the container with id 88a86f87627571f8098ea758f8461535421de9cb4d41b69c490d81caa6caa3ea Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.446874 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.632268 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh"] Dec 05 08:29:38 crc kubenswrapper[4876]: W1205 08:29:38.640202 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85c45836_2564_48e9_8c2f_33db89131c1b.slice/crio-b43f5352992a9f664b3ddb889ef0d34909d1537e8b5e856fcd3a228f761e26c1 WatchSource:0}: Error finding container b43f5352992a9f664b3ddb889ef0d34909d1537e8b5e856fcd3a228f761e26c1: Status 404 returned error can't find the container with id b43f5352992a9f664b3ddb889ef0d34909d1537e8b5e856fcd3a228f761e26c1 Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.758107 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" event={"ID":"85c45836-2564-48e9-8c2f-33db89131c1b","Type":"ContainerStarted","Data":"b43f5352992a9f664b3ddb889ef0d34909d1537e8b5e856fcd3a228f761e26c1"} Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.759568 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" event={"ID":"2fb0c525-e24c-4093-b8f7-8a9e3f788c29","Type":"ContainerStarted","Data":"f427b5c49ccb15892f9dd08899f0184db038a79726fcb9ad9fe7c70aed6d4333"} Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.761479 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-576575b889-kmlg4" event={"ID":"658d8a8a-3c7c-4e64-a98c-2f619b52c21e","Type":"ContainerStarted","Data":"00d5493f759f4d9de1f20ad696aa13a43642a7cee2fe026559cbb8d0a5152c54"} Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.761503 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-576575b889-kmlg4" event={"ID":"658d8a8a-3c7c-4e64-a98c-2f619b52c21e","Type":"ContainerStarted","Data":"88a86f87627571f8098ea758f8461535421de9cb4d41b69c490d81caa6caa3ea"} Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.762330 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-n2snz" event={"ID":"8c0975c5-f6cd-4565-a5fc-c9247bc8871c","Type":"ContainerStarted","Data":"ee62cd1259e7adc3faca2f43459ce9643e90494dd47065c623993c9221a021ba"} Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.765087 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" event={"ID":"f256549a-a4ac-40d4-b0b9-74e5b888d9f4","Type":"ContainerStarted","Data":"f38788a253f9fc5aecc6ab74c0102b220bda79687c865d45d30d9b7a3f95710c"} Dec 05 08:29:38 crc kubenswrapper[4876]: I1205 08:29:38.786598 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-576575b889-kmlg4" podStartSLOduration=1.786578904 podStartE2EDuration="1.786578904s" podCreationTimestamp="2025-12-05 08:29:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:29:38.785299149 +0000 UTC m=+843.273963761" watchObservedRunningTime="2025-12-05 08:29:38.786578904 +0000 UTC m=+843.275243546" Dec 05 08:29:39 crc kubenswrapper[4876]: I1205 08:29:39.916432 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:39 crc kubenswrapper[4876]: I1205 08:29:39.964651 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:40 crc kubenswrapper[4876]: I1205 08:29:40.149068 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-plkcw"] Dec 05 08:29:41 crc kubenswrapper[4876]: I1205 08:29:41.785543 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-plkcw" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="registry-server" containerID="cri-o://ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79" gracePeriod=2 Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.282873 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.387475 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpccz\" (UniqueName: \"kubernetes.io/projected/96310d2c-83ba-4695-afaa-6a9e2490b066-kube-api-access-bpccz\") pod \"96310d2c-83ba-4695-afaa-6a9e2490b066\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.387528 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-utilities\") pod \"96310d2c-83ba-4695-afaa-6a9e2490b066\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.387645 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-catalog-content\") pod \"96310d2c-83ba-4695-afaa-6a9e2490b066\" (UID: \"96310d2c-83ba-4695-afaa-6a9e2490b066\") " Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.388634 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-utilities" (OuterVolumeSpecName: "utilities") pod "96310d2c-83ba-4695-afaa-6a9e2490b066" (UID: "96310d2c-83ba-4695-afaa-6a9e2490b066"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.392863 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96310d2c-83ba-4695-afaa-6a9e2490b066-kube-api-access-bpccz" (OuterVolumeSpecName: "kube-api-access-bpccz") pod "96310d2c-83ba-4695-afaa-6a9e2490b066" (UID: "96310d2c-83ba-4695-afaa-6a9e2490b066"). InnerVolumeSpecName "kube-api-access-bpccz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.488796 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpccz\" (UniqueName: \"kubernetes.io/projected/96310d2c-83ba-4695-afaa-6a9e2490b066-kube-api-access-bpccz\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.488832 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.507001 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96310d2c-83ba-4695-afaa-6a9e2490b066" (UID: "96310d2c-83ba-4695-afaa-6a9e2490b066"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.589926 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96310d2c-83ba-4695-afaa-6a9e2490b066-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.792597 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" event={"ID":"2fb0c525-e24c-4093-b8f7-8a9e3f788c29","Type":"ContainerStarted","Data":"014182e68a66e075eac1e4463db69c1d459bff045b5e5a428ee1f17fd7470797"} Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.795408 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" event={"ID":"85c45836-2564-48e9-8c2f-33db89131c1b","Type":"ContainerStarted","Data":"99e720dc6e9aa6a71a9cfbd6d16527d1710b64eeaa1867bf4c314043f1637108"} Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.798350 4876 generic.go:334] "Generic (PLEG): container finished" podID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerID="ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79" exitCode=0 Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.798503 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-plkcw" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.798840 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plkcw" event={"ID":"96310d2c-83ba-4695-afaa-6a9e2490b066","Type":"ContainerDied","Data":"ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79"} Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.798870 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plkcw" event={"ID":"96310d2c-83ba-4695-afaa-6a9e2490b066","Type":"ContainerDied","Data":"09d7e259b51fd7ea1595bf8df0306a033c81935a24375868eb28ab3962c220a6"} Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.798889 4876 scope.go:117] "RemoveContainer" containerID="ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.801553 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-n2snz" event={"ID":"8c0975c5-f6cd-4565-a5fc-c9247bc8871c","Type":"ContainerStarted","Data":"9e2a923a1f311a522fee50cda6023df6f8ac3149ace7c6532671b31801aa931f"} Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.801707 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.803185 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" event={"ID":"f256549a-a4ac-40d4-b0b9-74e5b888d9f4","Type":"ContainerStarted","Data":"8cb3b0fb38bd69bfdc7c13c049fb3d4914072b5a29a5de9ff620d4c42b81fbd2"} Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.803496 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.811559 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g4pnh" podStartSLOduration=2.4643516 podStartE2EDuration="5.811543119s" podCreationTimestamp="2025-12-05 08:29:37 +0000 UTC" firstStartedPulling="2025-12-05 08:29:38.642221153 +0000 UTC m=+843.130885775" lastFinishedPulling="2025-12-05 08:29:41.989412672 +0000 UTC m=+846.478077294" observedRunningTime="2025-12-05 08:29:42.810027177 +0000 UTC m=+847.298691809" watchObservedRunningTime="2025-12-05 08:29:42.811543119 +0000 UTC m=+847.300207761" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.824624 4876 scope.go:117] "RemoveContainer" containerID="07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.853803 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-n2snz" podStartSLOduration=1.854904034 podStartE2EDuration="5.853783031s" podCreationTimestamp="2025-12-05 08:29:37 +0000 UTC" firstStartedPulling="2025-12-05 08:29:38.026726961 +0000 UTC m=+842.515391603" lastFinishedPulling="2025-12-05 08:29:42.025605978 +0000 UTC m=+846.514270600" observedRunningTime="2025-12-05 08:29:42.83159078 +0000 UTC m=+847.320255412" watchObservedRunningTime="2025-12-05 08:29:42.853783031 +0000 UTC m=+847.342447653" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.857613 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" podStartSLOduration=1.9885706810000001 podStartE2EDuration="5.857596646s" podCreationTimestamp="2025-12-05 08:29:37 +0000 UTC" firstStartedPulling="2025-12-05 08:29:38.195181525 +0000 UTC m=+842.683846137" lastFinishedPulling="2025-12-05 08:29:42.06420748 +0000 UTC m=+846.552872102" observedRunningTime="2025-12-05 08:29:42.854548902 +0000 UTC m=+847.343213524" watchObservedRunningTime="2025-12-05 08:29:42.857596646 +0000 UTC m=+847.346261258" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.870251 4876 scope.go:117] "RemoveContainer" containerID="5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.875310 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-plkcw"] Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.879464 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-plkcw"] Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.898546 4876 scope.go:117] "RemoveContainer" containerID="ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79" Dec 05 08:29:42 crc kubenswrapper[4876]: E1205 08:29:42.898973 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79\": container with ID starting with ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79 not found: ID does not exist" containerID="ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.899077 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79"} err="failed to get container status \"ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79\": rpc error: code = NotFound desc = could not find container \"ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79\": container with ID starting with ac45592ef0e27772f37ef04bda5353e823a3343f9ae56044eb42db0ce2450a79 not found: ID does not exist" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.899134 4876 scope.go:117] "RemoveContainer" containerID="07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd" Dec 05 08:29:42 crc kubenswrapper[4876]: E1205 08:29:42.899584 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd\": container with ID starting with 07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd not found: ID does not exist" containerID="07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.899622 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd"} err="failed to get container status \"07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd\": rpc error: code = NotFound desc = could not find container \"07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd\": container with ID starting with 07092b94cf5a1ca0a462e154a3f18b44d6c6d0ae4b1b168036a45d0a47de09dd not found: ID does not exist" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.899649 4876 scope.go:117] "RemoveContainer" containerID="5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750" Dec 05 08:29:42 crc kubenswrapper[4876]: E1205 08:29:42.900083 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750\": container with ID starting with 5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750 not found: ID does not exist" containerID="5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750" Dec 05 08:29:42 crc kubenswrapper[4876]: I1205 08:29:42.900109 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750"} err="failed to get container status \"5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750\": rpc error: code = NotFound desc = could not find container \"5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750\": container with ID starting with 5215a8a6c2df2afa0c46e033d3e308739dbcf25be8249566283843e4cc05a750 not found: ID does not exist" Dec 05 08:29:43 crc kubenswrapper[4876]: I1205 08:29:43.831316 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" path="/var/lib/kubelet/pods/96310d2c-83ba-4695-afaa-6a9e2490b066/volumes" Dec 05 08:29:44 crc kubenswrapper[4876]: I1205 08:29:44.823533 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" event={"ID":"2fb0c525-e24c-4093-b8f7-8a9e3f788c29","Type":"ContainerStarted","Data":"0573f8fe4696660f2eb6b97b9ab129ccef307d2470407fb897f6593b8416783f"} Dec 05 08:29:44 crc kubenswrapper[4876]: I1205 08:29:44.851713 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n4rqm" podStartSLOduration=1.682907403 podStartE2EDuration="7.851675931s" podCreationTimestamp="2025-12-05 08:29:37 +0000 UTC" firstStartedPulling="2025-12-05 08:29:38.172304536 +0000 UTC m=+842.660969158" lastFinishedPulling="2025-12-05 08:29:44.341073064 +0000 UTC m=+848.829737686" observedRunningTime="2025-12-05 08:29:44.843460155 +0000 UTC m=+849.332124807" watchObservedRunningTime="2025-12-05 08:29:44.851675931 +0000 UTC m=+849.340340593" Dec 05 08:29:47 crc kubenswrapper[4876]: I1205 08:29:47.975769 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:47 crc kubenswrapper[4876]: I1205 08:29:47.976173 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:47 crc kubenswrapper[4876]: I1205 08:29:47.984651 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:48 crc kubenswrapper[4876]: I1205 08:29:48.043590 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-n2snz" Dec 05 08:29:48 crc kubenswrapper[4876]: I1205 08:29:48.857348 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-576575b889-kmlg4" Dec 05 08:29:48 crc kubenswrapper[4876]: I1205 08:29:48.985276 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-plzs7"] Dec 05 08:29:57 crc kubenswrapper[4876]: I1205 08:29:57.697882 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-dhgxp" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.147344 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm"] Dec 05 08:30:00 crc kubenswrapper[4876]: E1205 08:30:00.147949 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="registry-server" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.147968 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="registry-server" Dec 05 08:30:00 crc kubenswrapper[4876]: E1205 08:30:00.147981 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="extract-utilities" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.147989 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="extract-utilities" Dec 05 08:30:00 crc kubenswrapper[4876]: E1205 08:30:00.148010 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="extract-content" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.148018 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="extract-content" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.148143 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="96310d2c-83ba-4695-afaa-6a9e2490b066" containerName="registry-server" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.148613 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.150488 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.152891 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.157702 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm"] Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.339482 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz7g8\" (UniqueName: \"kubernetes.io/projected/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-kube-api-access-kz7g8\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.339560 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-secret-volume\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.339689 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-config-volume\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.441181 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz7g8\" (UniqueName: \"kubernetes.io/projected/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-kube-api-access-kz7g8\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.441260 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-secret-volume\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.441307 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-config-volume\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.442453 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-config-volume\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.448427 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-secret-volume\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.467279 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz7g8\" (UniqueName: \"kubernetes.io/projected/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-kube-api-access-kz7g8\") pod \"collect-profiles-29415390-hm9qm\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:00 crc kubenswrapper[4876]: I1205 08:30:00.766077 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:01 crc kubenswrapper[4876]: I1205 08:30:01.156524 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm"] Dec 05 08:30:01 crc kubenswrapper[4876]: W1205 08:30:01.161276 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aec6a5d_c8c9_4046_8fe8_b61d6345f0f9.slice/crio-8dc1aa9144f7fdffe6263ee9a85c4fb05351f5c1fe230a28109b6c30d3a56de2 WatchSource:0}: Error finding container 8dc1aa9144f7fdffe6263ee9a85c4fb05351f5c1fe230a28109b6c30d3a56de2: Status 404 returned error can't find the container with id 8dc1aa9144f7fdffe6263ee9a85c4fb05351f5c1fe230a28109b6c30d3a56de2 Dec 05 08:30:01 crc kubenswrapper[4876]: I1205 08:30:01.933324 4876 generic.go:334] "Generic (PLEG): container finished" podID="4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9" containerID="9bdc5e8bd2d68fceed99329241e11014fff9acb9a49aee40f89ff67098c3866a" exitCode=0 Dec 05 08:30:01 crc kubenswrapper[4876]: I1205 08:30:01.933457 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" event={"ID":"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9","Type":"ContainerDied","Data":"9bdc5e8bd2d68fceed99329241e11014fff9acb9a49aee40f89ff67098c3866a"} Dec 05 08:30:01 crc kubenswrapper[4876]: I1205 08:30:01.933585 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" event={"ID":"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9","Type":"ContainerStarted","Data":"8dc1aa9144f7fdffe6263ee9a85c4fb05351f5c1fe230a28109b6c30d3a56de2"} Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.144655 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.315296 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-secret-volume\") pod \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.315363 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz7g8\" (UniqueName: \"kubernetes.io/projected/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-kube-api-access-kz7g8\") pod \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.315450 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-config-volume\") pod \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\" (UID: \"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9\") " Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.316164 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-config-volume" (OuterVolumeSpecName: "config-volume") pod "4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9" (UID: "4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.320258 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-kube-api-access-kz7g8" (OuterVolumeSpecName: "kube-api-access-kz7g8") pod "4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9" (UID: "4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9"). InnerVolumeSpecName "kube-api-access-kz7g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.320292 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9" (UID: "4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.417568 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.417614 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz7g8\" (UniqueName: \"kubernetes.io/projected/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-kube-api-access-kz7g8\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.417629 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.946233 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" event={"ID":"4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9","Type":"ContainerDied","Data":"8dc1aa9144f7fdffe6263ee9a85c4fb05351f5c1fe230a28109b6c30d3a56de2"} Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.946643 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dc1aa9144f7fdffe6263ee9a85c4fb05351f5c1fe230a28109b6c30d3a56de2" Dec 05 08:30:03 crc kubenswrapper[4876]: I1205 08:30:03.946299 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm" Dec 05 08:30:08 crc kubenswrapper[4876]: I1205 08:30:08.213628 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:30:08 crc kubenswrapper[4876]: I1205 08:30:08.214263 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:30:10 crc kubenswrapper[4876]: I1205 08:30:10.935076 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd"] Dec 05 08:30:10 crc kubenswrapper[4876]: E1205 08:30:10.935984 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9" containerName="collect-profiles" Dec 05 08:30:10 crc kubenswrapper[4876]: I1205 08:30:10.936005 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9" containerName="collect-profiles" Dec 05 08:30:10 crc kubenswrapper[4876]: I1205 08:30:10.936170 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9" containerName="collect-profiles" Dec 05 08:30:10 crc kubenswrapper[4876]: I1205 08:30:10.937375 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:10 crc kubenswrapper[4876]: I1205 08:30:10.940328 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 08:30:10 crc kubenswrapper[4876]: I1205 08:30:10.951259 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd"] Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.118564 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5qh6\" (UniqueName: \"kubernetes.io/projected/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-kube-api-access-w5qh6\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.118779 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.118813 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.219774 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.219864 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.220009 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5qh6\" (UniqueName: \"kubernetes.io/projected/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-kube-api-access-w5qh6\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.220663 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.220743 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.241122 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5qh6\" (UniqueName: \"kubernetes.io/projected/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-kube-api-access-w5qh6\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.292120 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.723177 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd"] Dec 05 08:30:11 crc kubenswrapper[4876]: I1205 08:30:11.995189 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" event={"ID":"41d46d23-6d8e-44e3-ba8e-9be76b08bff6","Type":"ContainerStarted","Data":"76ce76d8bf7db0382c15e0a3bdb46cf0c5d4d4f7e7e67d12fcd9f86197731654"} Dec 05 08:30:13 crc kubenswrapper[4876]: I1205 08:30:13.004150 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" event={"ID":"41d46d23-6d8e-44e3-ba8e-9be76b08bff6","Type":"ContainerStarted","Data":"b37292cb36cdb07ffb1425d80b3153bb27ca5741f4c3ea383413db3a8b720b31"} Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.011007 4876 generic.go:334] "Generic (PLEG): container finished" podID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerID="b37292cb36cdb07ffb1425d80b3153bb27ca5741f4c3ea383413db3a8b720b31" exitCode=0 Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.011059 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" event={"ID":"41d46d23-6d8e-44e3-ba8e-9be76b08bff6","Type":"ContainerDied","Data":"b37292cb36cdb07ffb1425d80b3153bb27ca5741f4c3ea383413db3a8b720b31"} Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.049582 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-plzs7" podUID="75f4a218-73a6-4a91-a8ab-6177ee8d1e45" containerName="console" containerID="cri-o://223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c" gracePeriod=15 Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.394378 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-plzs7_75f4a218-73a6-4a91-a8ab-6177ee8d1e45/console/0.log" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.394752 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.572519 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-oauth-serving-cert\") pod \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.572563 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-config\") pod \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.572584 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-service-ca\") pod \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.572628 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-oauth-config\") pod \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.572661 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-serving-cert\") pod \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.572681 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-trusted-ca-bundle\") pod \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.572748 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrzmz\" (UniqueName: \"kubernetes.io/projected/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-kube-api-access-wrzmz\") pod \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\" (UID: \"75f4a218-73a6-4a91-a8ab-6177ee8d1e45\") " Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.573428 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "75f4a218-73a6-4a91-a8ab-6177ee8d1e45" (UID: "75f4a218-73a6-4a91-a8ab-6177ee8d1e45"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.573415 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-config" (OuterVolumeSpecName: "console-config") pod "75f4a218-73a6-4a91-a8ab-6177ee8d1e45" (UID: "75f4a218-73a6-4a91-a8ab-6177ee8d1e45"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.573458 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-service-ca" (OuterVolumeSpecName: "service-ca") pod "75f4a218-73a6-4a91-a8ab-6177ee8d1e45" (UID: "75f4a218-73a6-4a91-a8ab-6177ee8d1e45"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.574092 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "75f4a218-73a6-4a91-a8ab-6177ee8d1e45" (UID: "75f4a218-73a6-4a91-a8ab-6177ee8d1e45"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.592147 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "75f4a218-73a6-4a91-a8ab-6177ee8d1e45" (UID: "75f4a218-73a6-4a91-a8ab-6177ee8d1e45"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.592198 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-kube-api-access-wrzmz" (OuterVolumeSpecName: "kube-api-access-wrzmz") pod "75f4a218-73a6-4a91-a8ab-6177ee8d1e45" (UID: "75f4a218-73a6-4a91-a8ab-6177ee8d1e45"). InnerVolumeSpecName "kube-api-access-wrzmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.593017 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "75f4a218-73a6-4a91-a8ab-6177ee8d1e45" (UID: "75f4a218-73a6-4a91-a8ab-6177ee8d1e45"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.674095 4876 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.674151 4876 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.674167 4876 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.674180 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrzmz\" (UniqueName: \"kubernetes.io/projected/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-kube-api-access-wrzmz\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.674190 4876 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.674201 4876 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:14 crc kubenswrapper[4876]: I1205 08:30:14.674210 4876 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/75f4a218-73a6-4a91-a8ab-6177ee8d1e45-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.021431 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-plzs7_75f4a218-73a6-4a91-a8ab-6177ee8d1e45/console/0.log" Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.021495 4876 generic.go:334] "Generic (PLEG): container finished" podID="75f4a218-73a6-4a91-a8ab-6177ee8d1e45" containerID="223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c" exitCode=2 Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.021535 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-plzs7" event={"ID":"75f4a218-73a6-4a91-a8ab-6177ee8d1e45","Type":"ContainerDied","Data":"223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c"} Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.021566 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-plzs7" event={"ID":"75f4a218-73a6-4a91-a8ab-6177ee8d1e45","Type":"ContainerDied","Data":"18860c21b3f9fe8edfdd6d0decf3e6b1c728cfe43626cbb36a6fed4c6f20c660"} Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.021592 4876 scope.go:117] "RemoveContainer" containerID="223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c" Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.021604 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-plzs7" Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.043070 4876 scope.go:117] "RemoveContainer" containerID="223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c" Dec 05 08:30:15 crc kubenswrapper[4876]: E1205 08:30:15.043734 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c\": container with ID starting with 223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c not found: ID does not exist" containerID="223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c" Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.043800 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c"} err="failed to get container status \"223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c\": rpc error: code = NotFound desc = could not find container \"223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c\": container with ID starting with 223ffdac36b1f3f34170180a441b23aa29b1189d41437ddebb3529f7bad7536c not found: ID does not exist" Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.068006 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-plzs7"] Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.071090 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-plzs7"] Dec 05 08:30:15 crc kubenswrapper[4876]: I1205 08:30:15.837053 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75f4a218-73a6-4a91-a8ab-6177ee8d1e45" path="/var/lib/kubelet/pods/75f4a218-73a6-4a91-a8ab-6177ee8d1e45/volumes" Dec 05 08:30:16 crc kubenswrapper[4876]: I1205 08:30:16.060020 4876 generic.go:334] "Generic (PLEG): container finished" podID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerID="fc7c9d38967106b391ec2e3bf8ea5f08881e66f324940594b2a9b93b88b8512e" exitCode=0 Dec 05 08:30:16 crc kubenswrapper[4876]: I1205 08:30:16.060071 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" event={"ID":"41d46d23-6d8e-44e3-ba8e-9be76b08bff6","Type":"ContainerDied","Data":"fc7c9d38967106b391ec2e3bf8ea5f08881e66f324940594b2a9b93b88b8512e"} Dec 05 08:30:17 crc kubenswrapper[4876]: I1205 08:30:17.077108 4876 generic.go:334] "Generic (PLEG): container finished" podID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerID="62a40f8642331510bd3803543b00e093c1c0fc8c740e4d9a4f293ab24d2c5254" exitCode=0 Dec 05 08:30:17 crc kubenswrapper[4876]: I1205 08:30:17.077233 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" event={"ID":"41d46d23-6d8e-44e3-ba8e-9be76b08bff6","Type":"ContainerDied","Data":"62a40f8642331510bd3803543b00e093c1c0fc8c740e4d9a4f293ab24d2c5254"} Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.316370 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.421766 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-bundle\") pod \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.421986 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5qh6\" (UniqueName: \"kubernetes.io/projected/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-kube-api-access-w5qh6\") pod \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.422113 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-util\") pod \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\" (UID: \"41d46d23-6d8e-44e3-ba8e-9be76b08bff6\") " Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.422840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-bundle" (OuterVolumeSpecName: "bundle") pod "41d46d23-6d8e-44e3-ba8e-9be76b08bff6" (UID: "41d46d23-6d8e-44e3-ba8e-9be76b08bff6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.429175 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-kube-api-access-w5qh6" (OuterVolumeSpecName: "kube-api-access-w5qh6") pod "41d46d23-6d8e-44e3-ba8e-9be76b08bff6" (UID: "41d46d23-6d8e-44e3-ba8e-9be76b08bff6"). InnerVolumeSpecName "kube-api-access-w5qh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.437599 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-util" (OuterVolumeSpecName: "util") pod "41d46d23-6d8e-44e3-ba8e-9be76b08bff6" (UID: "41d46d23-6d8e-44e3-ba8e-9be76b08bff6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.523525 4876 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-util\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.523568 4876 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:18 crc kubenswrapper[4876]: I1205 08:30:18.523581 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5qh6\" (UniqueName: \"kubernetes.io/projected/41d46d23-6d8e-44e3-ba8e-9be76b08bff6-kube-api-access-w5qh6\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:19 crc kubenswrapper[4876]: I1205 08:30:19.095825 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" event={"ID":"41d46d23-6d8e-44e3-ba8e-9be76b08bff6","Type":"ContainerDied","Data":"76ce76d8bf7db0382c15e0a3bdb46cf0c5d4d4f7e7e67d12fcd9f86197731654"} Dec 05 08:30:19 crc kubenswrapper[4876]: I1205 08:30:19.095978 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76ce76d8bf7db0382c15e0a3bdb46cf0c5d4d4f7e7e67d12fcd9f86197731654" Dec 05 08:30:19 crc kubenswrapper[4876]: I1205 08:30:19.095992 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.273917 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5gg6n"] Dec 05 08:30:27 crc kubenswrapper[4876]: E1205 08:30:27.274631 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerName="pull" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.274644 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerName="pull" Dec 05 08:30:27 crc kubenswrapper[4876]: E1205 08:30:27.274663 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerName="extract" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.274669 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerName="extract" Dec 05 08:30:27 crc kubenswrapper[4876]: E1205 08:30:27.274681 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerName="util" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.274687 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerName="util" Dec 05 08:30:27 crc kubenswrapper[4876]: E1205 08:30:27.274696 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f4a218-73a6-4a91-a8ab-6177ee8d1e45" containerName="console" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.274701 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f4a218-73a6-4a91-a8ab-6177ee8d1e45" containerName="console" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.274786 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f4a218-73a6-4a91-a8ab-6177ee8d1e45" containerName="console" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.274802 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d46d23-6d8e-44e3-ba8e-9be76b08bff6" containerName="extract" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.275489 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.282230 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5gg6n"] Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.436490 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-utilities\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.436562 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78jz9\" (UniqueName: \"kubernetes.io/projected/42e3f645-ba5e-4717-acab-a595a30d1185-kube-api-access-78jz9\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.436624 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-catalog-content\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.537882 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-utilities\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.538320 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78jz9\" (UniqueName: \"kubernetes.io/projected/42e3f645-ba5e-4717-acab-a595a30d1185-kube-api-access-78jz9\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.538332 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-utilities\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.538410 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-catalog-content\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.538750 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-catalog-content\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.556032 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78jz9\" (UniqueName: \"kubernetes.io/projected/42e3f645-ba5e-4717-acab-a595a30d1185-kube-api-access-78jz9\") pod \"certified-operators-5gg6n\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.592589 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:27 crc kubenswrapper[4876]: I1205 08:30:27.810205 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5gg6n"] Dec 05 08:30:28 crc kubenswrapper[4876]: I1205 08:30:28.145117 4876 generic.go:334] "Generic (PLEG): container finished" podID="42e3f645-ba5e-4717-acab-a595a30d1185" containerID="a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2" exitCode=0 Dec 05 08:30:28 crc kubenswrapper[4876]: I1205 08:30:28.145177 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg6n" event={"ID":"42e3f645-ba5e-4717-acab-a595a30d1185","Type":"ContainerDied","Data":"a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2"} Dec 05 08:30:28 crc kubenswrapper[4876]: I1205 08:30:28.145203 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg6n" event={"ID":"42e3f645-ba5e-4717-acab-a595a30d1185","Type":"ContainerStarted","Data":"1d08e328ff6a95637639e3de75eafda33e45fbaaf6875724f9c2420558343bdf"} Dec 05 08:30:29 crc kubenswrapper[4876]: I1205 08:30:29.170389 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg6n" event={"ID":"42e3f645-ba5e-4717-acab-a595a30d1185","Type":"ContainerStarted","Data":"5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163"} Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.152629 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp"] Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.153430 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.155593 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-llqlp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.155833 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.156003 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.158267 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.158574 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.177819 4876 generic.go:334] "Generic (PLEG): container finished" podID="42e3f645-ba5e-4717-acab-a595a30d1185" containerID="5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163" exitCode=0 Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.178054 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg6n" event={"ID":"42e3f645-ba5e-4717-acab-a595a30d1185","Type":"ContainerDied","Data":"5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163"} Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.236696 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp"] Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.273002 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aea74793-6a5c-4aab-af14-8029af652439-apiservice-cert\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.273068 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwch9\" (UniqueName: \"kubernetes.io/projected/aea74793-6a5c-4aab-af14-8029af652439-kube-api-access-zwch9\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.273177 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aea74793-6a5c-4aab-af14-8029af652439-webhook-cert\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.374456 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aea74793-6a5c-4aab-af14-8029af652439-apiservice-cert\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.374498 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwch9\" (UniqueName: \"kubernetes.io/projected/aea74793-6a5c-4aab-af14-8029af652439-kube-api-access-zwch9\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.374571 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aea74793-6a5c-4aab-af14-8029af652439-webhook-cert\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.380440 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aea74793-6a5c-4aab-af14-8029af652439-webhook-cert\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.385346 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aea74793-6a5c-4aab-af14-8029af652439-apiservice-cert\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.393663 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwch9\" (UniqueName: \"kubernetes.io/projected/aea74793-6a5c-4aab-af14-8029af652439-kube-api-access-zwch9\") pod \"metallb-operator-controller-manager-649974fd96-8cnpp\" (UID: \"aea74793-6a5c-4aab-af14-8029af652439\") " pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.472580 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.509156 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j"] Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.509757 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.511877 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.512140 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.512394 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-q4sdd" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.529949 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j"] Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.683227 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0710617c-73f2-44c9-a6fb-15f19c3a4aed-apiservice-cert\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.683606 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0710617c-73f2-44c9-a6fb-15f19c3a4aed-webhook-cert\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.683680 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cts9\" (UniqueName: \"kubernetes.io/projected/0710617c-73f2-44c9-a6fb-15f19c3a4aed-kube-api-access-6cts9\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.784505 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0710617c-73f2-44c9-a6fb-15f19c3a4aed-apiservice-cert\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.784548 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0710617c-73f2-44c9-a6fb-15f19c3a4aed-webhook-cert\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.784582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cts9\" (UniqueName: \"kubernetes.io/projected/0710617c-73f2-44c9-a6fb-15f19c3a4aed-kube-api-access-6cts9\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.790323 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0710617c-73f2-44c9-a6fb-15f19c3a4aed-webhook-cert\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.791798 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0710617c-73f2-44c9-a6fb-15f19c3a4aed-apiservice-cert\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.800090 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cts9\" (UniqueName: \"kubernetes.io/projected/0710617c-73f2-44c9-a6fb-15f19c3a4aed-kube-api-access-6cts9\") pod \"metallb-operator-webhook-server-5b5765d7c-vzz6j\" (UID: \"0710617c-73f2-44c9-a6fb-15f19c3a4aed\") " pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.865061 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:30 crc kubenswrapper[4876]: I1205 08:30:30.982630 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp"] Dec 05 08:30:31 crc kubenswrapper[4876]: W1205 08:30:31.055516 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaea74793_6a5c_4aab_af14_8029af652439.slice/crio-ece339f54ce954c43119f4055407d69f55bfecb320d9fc364a265b84bb07c464 WatchSource:0}: Error finding container ece339f54ce954c43119f4055407d69f55bfecb320d9fc364a265b84bb07c464: Status 404 returned error can't find the container with id ece339f54ce954c43119f4055407d69f55bfecb320d9fc364a265b84bb07c464 Dec 05 08:30:31 crc kubenswrapper[4876]: I1205 08:30:31.196568 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" event={"ID":"aea74793-6a5c-4aab-af14-8029af652439","Type":"ContainerStarted","Data":"ece339f54ce954c43119f4055407d69f55bfecb320d9fc364a265b84bb07c464"} Dec 05 08:30:31 crc kubenswrapper[4876]: I1205 08:30:31.201806 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg6n" event={"ID":"42e3f645-ba5e-4717-acab-a595a30d1185","Type":"ContainerStarted","Data":"a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15"} Dec 05 08:30:31 crc kubenswrapper[4876]: I1205 08:30:31.221611 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5gg6n" podStartSLOduration=1.6771257469999998 podStartE2EDuration="4.221594874s" podCreationTimestamp="2025-12-05 08:30:27 +0000 UTC" firstStartedPulling="2025-12-05 08:30:28.146566252 +0000 UTC m=+892.635230874" lastFinishedPulling="2025-12-05 08:30:30.691035379 +0000 UTC m=+895.179700001" observedRunningTime="2025-12-05 08:30:31.218070517 +0000 UTC m=+895.706735149" watchObservedRunningTime="2025-12-05 08:30:31.221594874 +0000 UTC m=+895.710259496" Dec 05 08:30:31 crc kubenswrapper[4876]: I1205 08:30:31.279708 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j"] Dec 05 08:30:31 crc kubenswrapper[4876]: W1205 08:30:31.282849 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0710617c_73f2_44c9_a6fb_15f19c3a4aed.slice/crio-26906a89a5d6086799dbf6a90bd9c1769c8df88de5d3a81d7c75134f40524410 WatchSource:0}: Error finding container 26906a89a5d6086799dbf6a90bd9c1769c8df88de5d3a81d7c75134f40524410: Status 404 returned error can't find the container with id 26906a89a5d6086799dbf6a90bd9c1769c8df88de5d3a81d7c75134f40524410 Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.209309 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" event={"ID":"0710617c-73f2-44c9-a6fb-15f19c3a4aed","Type":"ContainerStarted","Data":"26906a89a5d6086799dbf6a90bd9c1769c8df88de5d3a81d7c75134f40524410"} Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.680339 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dzr9q"] Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.681367 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.692191 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzr9q"] Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.709043 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtxdt\" (UniqueName: \"kubernetes.io/projected/a882b1a9-ad71-4cbe-b320-137a4d08a860-kube-api-access-dtxdt\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.709116 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-catalog-content\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.711983 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-utilities\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.812856 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtxdt\" (UniqueName: \"kubernetes.io/projected/a882b1a9-ad71-4cbe-b320-137a4d08a860-kube-api-access-dtxdt\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.812945 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-catalog-content\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.813011 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-utilities\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.813576 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-utilities\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.813694 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-catalog-content\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:32 crc kubenswrapper[4876]: I1205 08:30:32.835764 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtxdt\" (UniqueName: \"kubernetes.io/projected/a882b1a9-ad71-4cbe-b320-137a4d08a860-kube-api-access-dtxdt\") pod \"redhat-marketplace-dzr9q\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:33 crc kubenswrapper[4876]: I1205 08:30:33.022661 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:33 crc kubenswrapper[4876]: I1205 08:30:33.348957 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzr9q"] Dec 05 08:30:34 crc kubenswrapper[4876]: I1205 08:30:34.265256 4876 generic.go:334] "Generic (PLEG): container finished" podID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerID="98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab" exitCode=0 Dec 05 08:30:34 crc kubenswrapper[4876]: I1205 08:30:34.265535 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzr9q" event={"ID":"a882b1a9-ad71-4cbe-b320-137a4d08a860","Type":"ContainerDied","Data":"98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab"} Dec 05 08:30:34 crc kubenswrapper[4876]: I1205 08:30:34.265566 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzr9q" event={"ID":"a882b1a9-ad71-4cbe-b320-137a4d08a860","Type":"ContainerStarted","Data":"0783e1afdba05f29b0a5c941f09ecbfe236150a86a2d1a9aa510576472feb762"} Dec 05 08:30:37 crc kubenswrapper[4876]: I1205 08:30:37.593401 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:37 crc kubenswrapper[4876]: I1205 08:30:37.593868 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:37 crc kubenswrapper[4876]: I1205 08:30:37.650078 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.213167 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.213503 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.291434 4876 generic.go:334] "Generic (PLEG): container finished" podID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerID="ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39" exitCode=0 Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.291493 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzr9q" event={"ID":"a882b1a9-ad71-4cbe-b320-137a4d08a860","Type":"ContainerDied","Data":"ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39"} Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.293601 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" event={"ID":"0710617c-73f2-44c9-a6fb-15f19c3a4aed","Type":"ContainerStarted","Data":"dcb7ea402c248c0f50fc9d2916496cafc71c07fae97b0beff127a7add4777723"} Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.293727 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.296066 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" event={"ID":"aea74793-6a5c-4aab-af14-8029af652439","Type":"ContainerStarted","Data":"880bdfd8f3f2f90c52ca7e6097409f6e1a5a126824b9cafdfe6e774c80dfd84d"} Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.296314 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.346178 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" podStartSLOduration=1.935282577 podStartE2EDuration="8.346159684s" podCreationTimestamp="2025-12-05 08:30:30 +0000 UTC" firstStartedPulling="2025-12-05 08:30:31.285399609 +0000 UTC m=+895.774064231" lastFinishedPulling="2025-12-05 08:30:37.696276706 +0000 UTC m=+902.184941338" observedRunningTime="2025-12-05 08:30:38.345684171 +0000 UTC m=+902.834348793" watchObservedRunningTime="2025-12-05 08:30:38.346159684 +0000 UTC m=+902.834824306" Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.357509 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:38 crc kubenswrapper[4876]: I1205 08:30:38.388011 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" podStartSLOduration=1.80379603 podStartE2EDuration="8.387995825s" podCreationTimestamp="2025-12-05 08:30:30 +0000 UTC" firstStartedPulling="2025-12-05 08:30:31.059416513 +0000 UTC m=+895.548081135" lastFinishedPulling="2025-12-05 08:30:37.643616298 +0000 UTC m=+902.132280930" observedRunningTime="2025-12-05 08:30:38.368376535 +0000 UTC m=+902.857041157" watchObservedRunningTime="2025-12-05 08:30:38.387995825 +0000 UTC m=+902.876660447" Dec 05 08:30:39 crc kubenswrapper[4876]: I1205 08:30:39.303672 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzr9q" event={"ID":"a882b1a9-ad71-4cbe-b320-137a4d08a860","Type":"ContainerStarted","Data":"b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e"} Dec 05 08:30:39 crc kubenswrapper[4876]: I1205 08:30:39.323476 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dzr9q" podStartSLOduration=4.098848623 podStartE2EDuration="7.323434679s" podCreationTimestamp="2025-12-05 08:30:32 +0000 UTC" firstStartedPulling="2025-12-05 08:30:35.48185486 +0000 UTC m=+899.970519482" lastFinishedPulling="2025-12-05 08:30:38.706440916 +0000 UTC m=+903.195105538" observedRunningTime="2025-12-05 08:30:39.31913115 +0000 UTC m=+903.807795782" watchObservedRunningTime="2025-12-05 08:30:39.323434679 +0000 UTC m=+903.812099301" Dec 05 08:30:41 crc kubenswrapper[4876]: I1205 08:30:41.072711 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5gg6n"] Dec 05 08:30:41 crc kubenswrapper[4876]: I1205 08:30:41.073062 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5gg6n" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" containerName="registry-server" containerID="cri-o://a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15" gracePeriod=2 Dec 05 08:30:41 crc kubenswrapper[4876]: I1205 08:30:41.963992 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.061443 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78jz9\" (UniqueName: \"kubernetes.io/projected/42e3f645-ba5e-4717-acab-a595a30d1185-kube-api-access-78jz9\") pod \"42e3f645-ba5e-4717-acab-a595a30d1185\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.061530 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-utilities\") pod \"42e3f645-ba5e-4717-acab-a595a30d1185\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.061669 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-catalog-content\") pod \"42e3f645-ba5e-4717-acab-a595a30d1185\" (UID: \"42e3f645-ba5e-4717-acab-a595a30d1185\") " Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.063621 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-utilities" (OuterVolumeSpecName: "utilities") pod "42e3f645-ba5e-4717-acab-a595a30d1185" (UID: "42e3f645-ba5e-4717-acab-a595a30d1185"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.082092 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42e3f645-ba5e-4717-acab-a595a30d1185-kube-api-access-78jz9" (OuterVolumeSpecName: "kube-api-access-78jz9") pod "42e3f645-ba5e-4717-acab-a595a30d1185" (UID: "42e3f645-ba5e-4717-acab-a595a30d1185"). InnerVolumeSpecName "kube-api-access-78jz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.117561 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42e3f645-ba5e-4717-acab-a595a30d1185" (UID: "42e3f645-ba5e-4717-acab-a595a30d1185"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.162766 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.162809 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78jz9\" (UniqueName: \"kubernetes.io/projected/42e3f645-ba5e-4717-acab-a595a30d1185-kube-api-access-78jz9\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.162821 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e3f645-ba5e-4717-acab-a595a30d1185-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.331100 4876 generic.go:334] "Generic (PLEG): container finished" podID="42e3f645-ba5e-4717-acab-a595a30d1185" containerID="a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15" exitCode=0 Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.331140 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg6n" event={"ID":"42e3f645-ba5e-4717-acab-a595a30d1185","Type":"ContainerDied","Data":"a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15"} Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.331531 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gg6n" event={"ID":"42e3f645-ba5e-4717-acab-a595a30d1185","Type":"ContainerDied","Data":"1d08e328ff6a95637639e3de75eafda33e45fbaaf6875724f9c2420558343bdf"} Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.331215 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gg6n" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.331560 4876 scope.go:117] "RemoveContainer" containerID="a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.361107 4876 scope.go:117] "RemoveContainer" containerID="5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.366047 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5gg6n"] Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.381593 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5gg6n"] Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.396920 4876 scope.go:117] "RemoveContainer" containerID="a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.409812 4876 scope.go:117] "RemoveContainer" containerID="a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15" Dec 05 08:30:42 crc kubenswrapper[4876]: E1205 08:30:42.410360 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15\": container with ID starting with a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15 not found: ID does not exist" containerID="a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.410414 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15"} err="failed to get container status \"a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15\": rpc error: code = NotFound desc = could not find container \"a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15\": container with ID starting with a291ab260138956d332bd5ffd74b62a85daf7bd7e93b8d8d4b7b672c45d47c15 not found: ID does not exist" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.410446 4876 scope.go:117] "RemoveContainer" containerID="5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163" Dec 05 08:30:42 crc kubenswrapper[4876]: E1205 08:30:42.410739 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163\": container with ID starting with 5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163 not found: ID does not exist" containerID="5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.410764 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163"} err="failed to get container status \"5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163\": rpc error: code = NotFound desc = could not find container \"5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163\": container with ID starting with 5b52ae0cdc62b1063f9dd59c27ff6d3ea9ffd5b393ebe40b9e6bdc351068c163 not found: ID does not exist" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.410782 4876 scope.go:117] "RemoveContainer" containerID="a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2" Dec 05 08:30:42 crc kubenswrapper[4876]: E1205 08:30:42.411046 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2\": container with ID starting with a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2 not found: ID does not exist" containerID="a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2" Dec 05 08:30:42 crc kubenswrapper[4876]: I1205 08:30:42.411076 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2"} err="failed to get container status \"a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2\": rpc error: code = NotFound desc = could not find container \"a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2\": container with ID starting with a38ba0e85382357534130737d54051d8321b4bc59b155680f217fd40f0e434c2 not found: ID does not exist" Dec 05 08:30:43 crc kubenswrapper[4876]: I1205 08:30:43.023158 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:43 crc kubenswrapper[4876]: I1205 08:30:43.023197 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:43 crc kubenswrapper[4876]: I1205 08:30:43.080965 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:43 crc kubenswrapper[4876]: I1205 08:30:43.390494 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:43 crc kubenswrapper[4876]: I1205 08:30:43.831140 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" path="/var/lib/kubelet/pods/42e3f645-ba5e-4717-acab-a595a30d1185/volumes" Dec 05 08:30:44 crc kubenswrapper[4876]: I1205 08:30:44.671187 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzr9q"] Dec 05 08:30:45 crc kubenswrapper[4876]: I1205 08:30:45.348986 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dzr9q" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerName="registry-server" containerID="cri-o://b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e" gracePeriod=2 Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.230785 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.319358 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-catalog-content\") pod \"a882b1a9-ad71-4cbe-b320-137a4d08a860\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.319410 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtxdt\" (UniqueName: \"kubernetes.io/projected/a882b1a9-ad71-4cbe-b320-137a4d08a860-kube-api-access-dtxdt\") pod \"a882b1a9-ad71-4cbe-b320-137a4d08a860\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.319530 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-utilities\") pod \"a882b1a9-ad71-4cbe-b320-137a4d08a860\" (UID: \"a882b1a9-ad71-4cbe-b320-137a4d08a860\") " Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.320302 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-utilities" (OuterVolumeSpecName: "utilities") pod "a882b1a9-ad71-4cbe-b320-137a4d08a860" (UID: "a882b1a9-ad71-4cbe-b320-137a4d08a860"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.337127 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a882b1a9-ad71-4cbe-b320-137a4d08a860-kube-api-access-dtxdt" (OuterVolumeSpecName: "kube-api-access-dtxdt") pod "a882b1a9-ad71-4cbe-b320-137a4d08a860" (UID: "a882b1a9-ad71-4cbe-b320-137a4d08a860"). InnerVolumeSpecName "kube-api-access-dtxdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.359268 4876 generic.go:334] "Generic (PLEG): container finished" podID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerID="b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e" exitCode=0 Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.359314 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzr9q" event={"ID":"a882b1a9-ad71-4cbe-b320-137a4d08a860","Type":"ContainerDied","Data":"b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e"} Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.359346 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzr9q" event={"ID":"a882b1a9-ad71-4cbe-b320-137a4d08a860","Type":"ContainerDied","Data":"0783e1afdba05f29b0a5c941f09ecbfe236150a86a2d1a9aa510576472feb762"} Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.359366 4876 scope.go:117] "RemoveContainer" containerID="b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.359488 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzr9q" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.363136 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a882b1a9-ad71-4cbe-b320-137a4d08a860" (UID: "a882b1a9-ad71-4cbe-b320-137a4d08a860"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.388137 4876 scope.go:117] "RemoveContainer" containerID="ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.403143 4876 scope.go:117] "RemoveContainer" containerID="98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.420639 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.420680 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a882b1a9-ad71-4cbe-b320-137a4d08a860-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.420694 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtxdt\" (UniqueName: \"kubernetes.io/projected/a882b1a9-ad71-4cbe-b320-137a4d08a860-kube-api-access-dtxdt\") on node \"crc\" DevicePath \"\"" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.425362 4876 scope.go:117] "RemoveContainer" containerID="b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e" Dec 05 08:30:46 crc kubenswrapper[4876]: E1205 08:30:46.425865 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e\": container with ID starting with b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e not found: ID does not exist" containerID="b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.425944 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e"} err="failed to get container status \"b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e\": rpc error: code = NotFound desc = could not find container \"b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e\": container with ID starting with b4a48d9d475bc001a2b1f36cc072878fbb0ec582296b88a858f00ad31b81223e not found: ID does not exist" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.425974 4876 scope.go:117] "RemoveContainer" containerID="ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39" Dec 05 08:30:46 crc kubenswrapper[4876]: E1205 08:30:46.426549 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39\": container with ID starting with ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39 not found: ID does not exist" containerID="ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.426591 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39"} err="failed to get container status \"ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39\": rpc error: code = NotFound desc = could not find container \"ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39\": container with ID starting with ae9bbb50504e21e9c108bee43d6c98ab2644f8266069c46332316fed64925b39 not found: ID does not exist" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.426618 4876 scope.go:117] "RemoveContainer" containerID="98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab" Dec 05 08:30:46 crc kubenswrapper[4876]: E1205 08:30:46.427026 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab\": container with ID starting with 98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab not found: ID does not exist" containerID="98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.427057 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab"} err="failed to get container status \"98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab\": rpc error: code = NotFound desc = could not find container \"98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab\": container with ID starting with 98fa9bbfaaac5806bd4a5b96b76f0129b34a30990be3b44b574f23aba1ffcbab not found: ID does not exist" Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.683817 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzr9q"] Dec 05 08:30:46 crc kubenswrapper[4876]: I1205 08:30:46.688236 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzr9q"] Dec 05 08:30:47 crc kubenswrapper[4876]: I1205 08:30:47.830503 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" path="/var/lib/kubelet/pods/a882b1a9-ad71-4cbe-b320-137a4d08a860/volumes" Dec 05 08:30:50 crc kubenswrapper[4876]: I1205 08:30:50.874534 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5b5765d7c-vzz6j" Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.213762 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.214283 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.214322 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.214815 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b93e463bd75bc474356eb877c940a3ca93130dc5e558c7d99bfe4997ecfbcedb"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.214872 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://b93e463bd75bc474356eb877c940a3ca93130dc5e558c7d99bfe4997ecfbcedb" gracePeriod=600 Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.501586 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="b93e463bd75bc474356eb877c940a3ca93130dc5e558c7d99bfe4997ecfbcedb" exitCode=0 Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.501768 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"b93e463bd75bc474356eb877c940a3ca93130dc5e558c7d99bfe4997ecfbcedb"} Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.501920 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"6167cbd18c38bbbdb4c177b56aa429711e7d69fa0ae0c347c0b78af72fe60dcb"} Dec 05 08:31:08 crc kubenswrapper[4876]: I1205 08:31:08.501945 4876 scope.go:117] "RemoveContainer" containerID="3323fa5b6329c4296cfe91ae288ce25651562ac0fb3157978c65f5ccba6cd828" Dec 05 08:31:10 crc kubenswrapper[4876]: I1205 08:31:10.477788 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-649974fd96-8cnpp" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269326 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-fgbfx"] Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.269557 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" containerName="extract-utilities" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269570 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" containerName="extract-utilities" Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.269582 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" containerName="registry-server" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269588 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" containerName="registry-server" Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.269597 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerName="registry-server" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269605 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerName="registry-server" Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.269614 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerName="extract-content" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269621 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerName="extract-content" Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.269630 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerName="extract-utilities" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269636 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerName="extract-utilities" Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.269644 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" containerName="extract-content" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269649 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" containerName="extract-content" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269750 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a882b1a9-ad71-4cbe-b320-137a4d08a860" containerName="registry-server" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.269762 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e3f645-ba5e-4717-acab-a595a30d1185" containerName="registry-server" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.271656 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.277176 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4"] Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.277870 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.279073 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.279774 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.279811 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.279972 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4cd2f" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.296837 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4"] Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339664 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-startup\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339731 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn6f9\" (UniqueName: \"kubernetes.io/projected/2cdf28f4-aee9-424f-81b7-36e131a81f30-kube-api-access-zn6f9\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339754 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-metrics\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339789 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-reloader\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339806 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-sockets\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339821 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28f9db86-d329-4fa3-b808-b85aabf3840d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-gvmp4\" (UID: \"28f9db86-d329-4fa3-b808-b85aabf3840d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339839 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlmnp\" (UniqueName: \"kubernetes.io/projected/28f9db86-d329-4fa3-b808-b85aabf3840d-kube-api-access-wlmnp\") pod \"frr-k8s-webhook-server-7fcb986d4-gvmp4\" (UID: \"28f9db86-d329-4fa3-b808-b85aabf3840d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339864 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-conf\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.339878 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2cdf28f4-aee9-424f-81b7-36e131a81f30-metrics-certs\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.363559 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9k4hh"] Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.364403 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.365989 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.366099 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.366212 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-p6nmz" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.367077 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.379218 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-77dp9"] Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.380289 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.394861 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.425846 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-77dp9"] Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.440831 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-reloader\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.440892 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-sockets\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.440938 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28f9db86-d329-4fa3-b808-b85aabf3840d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-gvmp4\" (UID: \"28f9db86-d329-4fa3-b808-b85aabf3840d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.440967 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlmnp\" (UniqueName: \"kubernetes.io/projected/28f9db86-d329-4fa3-b808-b85aabf3840d-kube-api-access-wlmnp\") pod \"frr-k8s-webhook-server-7fcb986d4-gvmp4\" (UID: \"28f9db86-d329-4fa3-b808-b85aabf3840d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441005 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-conf\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441026 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2cdf28f4-aee9-424f-81b7-36e131a81f30-metrics-certs\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441063 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-cert\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441089 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-startup\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441117 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441157 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn6f9\" (UniqueName: \"kubernetes.io/projected/2cdf28f4-aee9-424f-81b7-36e131a81f30-kube-api-access-zn6f9\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441180 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-metrics-certs\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441206 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-metrics\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441234 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9a572583-7629-4436-b956-4b5befcb75ca-metallb-excludel2\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441257 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-metrics-certs\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441278 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcmzw\" (UniqueName: \"kubernetes.io/projected/9a572583-7629-4436-b956-4b5befcb75ca-kube-api-access-dcmzw\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441308 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9xrg\" (UniqueName: \"kubernetes.io/projected/460def3f-57ee-475e-bf28-788a977ea803-kube-api-access-k9xrg\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.441518 4876 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.441571 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28f9db86-d329-4fa3-b808-b85aabf3840d-cert podName:28f9db86-d329-4fa3-b808-b85aabf3840d nodeName:}" failed. No retries permitted until 2025-12-05 08:31:11.941551356 +0000 UTC m=+936.430215978 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/28f9db86-d329-4fa3-b808-b85aabf3840d-cert") pod "frr-k8s-webhook-server-7fcb986d4-gvmp4" (UID: "28f9db86-d329-4fa3-b808-b85aabf3840d") : secret "frr-k8s-webhook-server-cert" not found Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441777 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-reloader\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.441776 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-sockets\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.442152 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-conf\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.442220 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2cdf28f4-aee9-424f-81b7-36e131a81f30-metrics\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.445672 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2cdf28f4-aee9-424f-81b7-36e131a81f30-frr-startup\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.457292 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2cdf28f4-aee9-424f-81b7-36e131a81f30-metrics-certs\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.460306 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlmnp\" (UniqueName: \"kubernetes.io/projected/28f9db86-d329-4fa3-b808-b85aabf3840d-kube-api-access-wlmnp\") pod \"frr-k8s-webhook-server-7fcb986d4-gvmp4\" (UID: \"28f9db86-d329-4fa3-b808-b85aabf3840d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.461947 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn6f9\" (UniqueName: \"kubernetes.io/projected/2cdf28f4-aee9-424f-81b7-36e131a81f30-kube-api-access-zn6f9\") pod \"frr-k8s-fgbfx\" (UID: \"2cdf28f4-aee9-424f-81b7-36e131a81f30\") " pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.542793 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-metrics-certs\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.542839 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcmzw\" (UniqueName: \"kubernetes.io/projected/9a572583-7629-4436-b956-4b5befcb75ca-kube-api-access-dcmzw\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.542864 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9xrg\" (UniqueName: \"kubernetes.io/projected/460def3f-57ee-475e-bf28-788a977ea803-kube-api-access-k9xrg\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.542972 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-cert\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.543010 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.543037 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-metrics-certs\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.543068 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9a572583-7629-4436-b956-4b5befcb75ca-metallb-excludel2\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.543206 4876 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.543311 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist podName:9a572583-7629-4436-b956-4b5befcb75ca nodeName:}" failed. No retries permitted until 2025-12-05 08:31:12.04327795 +0000 UTC m=+936.531942572 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist") pod "speaker-9k4hh" (UID: "9a572583-7629-4436-b956-4b5befcb75ca") : secret "metallb-memberlist" not found Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.543321 4876 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 05 08:31:11 crc kubenswrapper[4876]: E1205 08:31:11.543366 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-metrics-certs podName:460def3f-57ee-475e-bf28-788a977ea803 nodeName:}" failed. No retries permitted until 2025-12-05 08:31:12.043350162 +0000 UTC m=+936.532014784 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-metrics-certs") pod "controller-f8648f98b-77dp9" (UID: "460def3f-57ee-475e-bf28-788a977ea803") : secret "controller-certs-secret" not found Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.543840 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9a572583-7629-4436-b956-4b5befcb75ca-metallb-excludel2\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.546554 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-metrics-certs\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.548460 4876 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.558796 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-cert\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.561383 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcmzw\" (UniqueName: \"kubernetes.io/projected/9a572583-7629-4436-b956-4b5befcb75ca-kube-api-access-dcmzw\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.573061 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9xrg\" (UniqueName: \"kubernetes.io/projected/460def3f-57ee-475e-bf28-788a977ea803-kube-api-access-k9xrg\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.591962 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.948626 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28f9db86-d329-4fa3-b808-b85aabf3840d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-gvmp4\" (UID: \"28f9db86-d329-4fa3-b808-b85aabf3840d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:11 crc kubenswrapper[4876]: I1205 08:31:11.954070 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28f9db86-d329-4fa3-b808-b85aabf3840d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-gvmp4\" (UID: \"28f9db86-d329-4fa3-b808-b85aabf3840d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.050054 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.050130 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-metrics-certs\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:12 crc kubenswrapper[4876]: E1205 08:31:12.050248 4876 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 08:31:12 crc kubenswrapper[4876]: E1205 08:31:12.050353 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist podName:9a572583-7629-4436-b956-4b5befcb75ca nodeName:}" failed. No retries permitted until 2025-12-05 08:31:13.050327689 +0000 UTC m=+937.538992381 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist") pod "speaker-9k4hh" (UID: "9a572583-7629-4436-b956-4b5befcb75ca") : secret "metallb-memberlist" not found Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.053262 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/460def3f-57ee-475e-bf28-788a977ea803-metrics-certs\") pod \"controller-f8648f98b-77dp9\" (UID: \"460def3f-57ee-475e-bf28-788a977ea803\") " pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.198499 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.316491 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.421360 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4"] Dec 05 08:31:12 crc kubenswrapper[4876]: W1205 08:31:12.430147 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28f9db86_d329_4fa3_b808_b85aabf3840d.slice/crio-6b2e14f5d3077397bdb6875bd6f5e5080af525785ded1dafa755b9098d35e22d WatchSource:0}: Error finding container 6b2e14f5d3077397bdb6875bd6f5e5080af525785ded1dafa755b9098d35e22d: Status 404 returned error can't find the container with id 6b2e14f5d3077397bdb6875bd6f5e5080af525785ded1dafa755b9098d35e22d Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.527051 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" event={"ID":"28f9db86-d329-4fa3-b808-b85aabf3840d","Type":"ContainerStarted","Data":"6b2e14f5d3077397bdb6875bd6f5e5080af525785ded1dafa755b9098d35e22d"} Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.528291 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerStarted","Data":"0b12d44a6a89cf6a36a7309a9cd5e2ad51e225dec51f22e64c3990f04e79bd0d"} Dec 05 08:31:12 crc kubenswrapper[4876]: I1205 08:31:12.730333 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-77dp9"] Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.076383 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.082233 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9a572583-7629-4436-b956-4b5befcb75ca-memberlist\") pod \"speaker-9k4hh\" (UID: \"9a572583-7629-4436-b956-4b5befcb75ca\") " pod="metallb-system/speaker-9k4hh" Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.178032 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9k4hh" Dec 05 08:31:13 crc kubenswrapper[4876]: W1205 08:31:13.247454 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a572583_7629_4436_b956_4b5befcb75ca.slice/crio-1409eb976cc04ef0a2073fb02b97eb43ae92260e9fec78ecf9707fb9b638c7ff WatchSource:0}: Error finding container 1409eb976cc04ef0a2073fb02b97eb43ae92260e9fec78ecf9707fb9b638c7ff: Status 404 returned error can't find the container with id 1409eb976cc04ef0a2073fb02b97eb43ae92260e9fec78ecf9707fb9b638c7ff Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.535958 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-77dp9" event={"ID":"460def3f-57ee-475e-bf28-788a977ea803","Type":"ContainerStarted","Data":"c6087693c968076598b5947af466d576646f378d8cd31432d2e9bbec50de007e"} Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.536344 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-77dp9" event={"ID":"460def3f-57ee-475e-bf28-788a977ea803","Type":"ContainerStarted","Data":"44a854b858ffb48ba27f269643046f7c601e909c4dbd2e6c75afa03a88b18d70"} Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.536363 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.536376 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-77dp9" event={"ID":"460def3f-57ee-475e-bf28-788a977ea803","Type":"ContainerStarted","Data":"ca9ee2cbcd682fc475af2a7857341e59c2c3b803128a991fef821fa64e8f7089"} Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.538769 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9k4hh" event={"ID":"9a572583-7629-4436-b956-4b5befcb75ca","Type":"ContainerStarted","Data":"1409eb976cc04ef0a2073fb02b97eb43ae92260e9fec78ecf9707fb9b638c7ff"} Dec 05 08:31:13 crc kubenswrapper[4876]: I1205 08:31:13.563231 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-77dp9" podStartSLOduration=2.56319041 podStartE2EDuration="2.56319041s" podCreationTimestamp="2025-12-05 08:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:31:13.553358481 +0000 UTC m=+938.042023103" watchObservedRunningTime="2025-12-05 08:31:13.56319041 +0000 UTC m=+938.051855032" Dec 05 08:31:14 crc kubenswrapper[4876]: I1205 08:31:14.569923 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9k4hh" event={"ID":"9a572583-7629-4436-b956-4b5befcb75ca","Type":"ContainerStarted","Data":"93a059cc4d34bc00c26ce4ec9744efb8b0988f1b4be1a2cd096d320a04b9266c"} Dec 05 08:31:14 crc kubenswrapper[4876]: I1205 08:31:14.569955 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9k4hh" event={"ID":"9a572583-7629-4436-b956-4b5befcb75ca","Type":"ContainerStarted","Data":"715d8558f35e28b69446b6071d796cce754f007a3427f159df65458f4466dfcc"} Dec 05 08:31:14 crc kubenswrapper[4876]: I1205 08:31:14.569975 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9k4hh" Dec 05 08:31:15 crc kubenswrapper[4876]: I1205 08:31:15.852865 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9k4hh" podStartSLOduration=4.852848512 podStartE2EDuration="4.852848512s" podCreationTimestamp="2025-12-05 08:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:31:14.588034051 +0000 UTC m=+939.076698673" watchObservedRunningTime="2025-12-05 08:31:15.852848512 +0000 UTC m=+940.341513134" Dec 05 08:31:21 crc kubenswrapper[4876]: I1205 08:31:21.623313 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" event={"ID":"28f9db86-d329-4fa3-b808-b85aabf3840d","Type":"ContainerStarted","Data":"0855edfc0009856175fd78039647ca03402c0c934a548fd5f252b5a4a1435cb0"} Dec 05 08:31:21 crc kubenswrapper[4876]: I1205 08:31:21.623914 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:21 crc kubenswrapper[4876]: I1205 08:31:21.625163 4876 generic.go:334] "Generic (PLEG): container finished" podID="2cdf28f4-aee9-424f-81b7-36e131a81f30" containerID="96bcf083fb00e48937c6c6930b499d6a58641710626b7bc29bb24391aa59a0fb" exitCode=0 Dec 05 08:31:21 crc kubenswrapper[4876]: I1205 08:31:21.625193 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerDied","Data":"96bcf083fb00e48937c6c6930b499d6a58641710626b7bc29bb24391aa59a0fb"} Dec 05 08:31:21 crc kubenswrapper[4876]: I1205 08:31:21.647446 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" podStartSLOduration=2.828414842 podStartE2EDuration="10.64742674s" podCreationTimestamp="2025-12-05 08:31:11 +0000 UTC" firstStartedPulling="2025-12-05 08:31:12.431791028 +0000 UTC m=+936.920455650" lastFinishedPulling="2025-12-05 08:31:20.250802926 +0000 UTC m=+944.739467548" observedRunningTime="2025-12-05 08:31:21.641604207 +0000 UTC m=+946.130268829" watchObservedRunningTime="2025-12-05 08:31:21.64742674 +0000 UTC m=+946.136091372" Dec 05 08:31:22 crc kubenswrapper[4876]: I1205 08:31:22.321041 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-77dp9" Dec 05 08:31:22 crc kubenswrapper[4876]: I1205 08:31:22.631963 4876 generic.go:334] "Generic (PLEG): container finished" podID="2cdf28f4-aee9-424f-81b7-36e131a81f30" containerID="5ed890628dd9ad9cd0aede2d38ad20a1ff661a99edaa0f1598a496717ff5f28d" exitCode=0 Dec 05 08:31:22 crc kubenswrapper[4876]: I1205 08:31:22.632279 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerDied","Data":"5ed890628dd9ad9cd0aede2d38ad20a1ff661a99edaa0f1598a496717ff5f28d"} Dec 05 08:31:23 crc kubenswrapper[4876]: I1205 08:31:23.187428 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9k4hh" Dec 05 08:31:23 crc kubenswrapper[4876]: I1205 08:31:23.643081 4876 generic.go:334] "Generic (PLEG): container finished" podID="2cdf28f4-aee9-424f-81b7-36e131a81f30" containerID="d6b160ccfafd8fd849a9b45a0a1ee8c2ed3ed6e156fffcf95ffa4292a4d4c139" exitCode=0 Dec 05 08:31:23 crc kubenswrapper[4876]: I1205 08:31:23.643308 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerDied","Data":"d6b160ccfafd8fd849a9b45a0a1ee8c2ed3ed6e156fffcf95ffa4292a4d4c139"} Dec 05 08:31:24 crc kubenswrapper[4876]: I1205 08:31:24.654785 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerStarted","Data":"b5c4c8c4a00c5c0c0cc798b1d8f6ff7ea6c997b9740841f5638a39467684f7b8"} Dec 05 08:31:24 crc kubenswrapper[4876]: I1205 08:31:24.655197 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerStarted","Data":"a46b029bb2591498e7368908b15d9967ebcf0cc84f4f24afa80332b8c4311223"} Dec 05 08:31:24 crc kubenswrapper[4876]: I1205 08:31:24.655222 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerStarted","Data":"d163a15162ce68a37b78338c97a09501bf119b4b04654a0eb89d404c68fb2148"} Dec 05 08:31:25 crc kubenswrapper[4876]: I1205 08:31:25.672429 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerStarted","Data":"92508c1f7da78eca5b8e99bb58f8cab45c5736a4674caa7a66a20854fc0f1095"} Dec 05 08:31:25 crc kubenswrapper[4876]: I1205 08:31:25.672830 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerStarted","Data":"0ea211a29e38566bbca5237635c2521de1938231dabf94268e19cdf6d6c4c1b4"} Dec 05 08:31:25 crc kubenswrapper[4876]: I1205 08:31:25.672849 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:25 crc kubenswrapper[4876]: I1205 08:31:25.672862 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fgbfx" event={"ID":"2cdf28f4-aee9-424f-81b7-36e131a81f30","Type":"ContainerStarted","Data":"197a264d62ea5d279ebb566d184918aa266f112635f29ce93a9e3e37dce76028"} Dec 05 08:31:25 crc kubenswrapper[4876]: I1205 08:31:25.707852 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-fgbfx" podStartSLOduration=6.212846232 podStartE2EDuration="14.707834712s" podCreationTimestamp="2025-12-05 08:31:11 +0000 UTC" firstStartedPulling="2025-12-05 08:31:11.739007675 +0000 UTC m=+936.227672307" lastFinishedPulling="2025-12-05 08:31:20.233996175 +0000 UTC m=+944.722660787" observedRunningTime="2025-12-05 08:31:25.698968399 +0000 UTC m=+950.187633021" watchObservedRunningTime="2025-12-05 08:31:25.707834712 +0000 UTC m=+950.196499334" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.213648 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qjjkm"] Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.215155 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.218574 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.218816 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-57d4r" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.224357 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.234284 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qjjkm"] Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.276937 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw2c7\" (UniqueName: \"kubernetes.io/projected/437c91ca-0472-481b-b58f-40269f4d2c18-kube-api-access-mw2c7\") pod \"openstack-operator-index-qjjkm\" (UID: \"437c91ca-0472-481b-b58f-40269f4d2c18\") " pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.381401 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw2c7\" (UniqueName: \"kubernetes.io/projected/437c91ca-0472-481b-b58f-40269f4d2c18-kube-api-access-mw2c7\") pod \"openstack-operator-index-qjjkm\" (UID: \"437c91ca-0472-481b-b58f-40269f4d2c18\") " pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.406596 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw2c7\" (UniqueName: \"kubernetes.io/projected/437c91ca-0472-481b-b58f-40269f4d2c18-kube-api-access-mw2c7\") pod \"openstack-operator-index-qjjkm\" (UID: \"437c91ca-0472-481b-b58f-40269f4d2c18\") " pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.538767 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.592674 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:26 crc kubenswrapper[4876]: I1205 08:31:26.648677 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:27 crc kubenswrapper[4876]: W1205 08:31:27.513765 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod437c91ca_0472_481b_b58f_40269f4d2c18.slice/crio-77ccb5e69477a06707381bf2672e15b1db8058dc218bd8b2179b7e76f1ada53e WatchSource:0}: Error finding container 77ccb5e69477a06707381bf2672e15b1db8058dc218bd8b2179b7e76f1ada53e: Status 404 returned error can't find the container with id 77ccb5e69477a06707381bf2672e15b1db8058dc218bd8b2179b7e76f1ada53e Dec 05 08:31:27 crc kubenswrapper[4876]: I1205 08:31:27.516340 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qjjkm"] Dec 05 08:31:27 crc kubenswrapper[4876]: I1205 08:31:27.691197 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qjjkm" event={"ID":"437c91ca-0472-481b-b58f-40269f4d2c18","Type":"ContainerStarted","Data":"77ccb5e69477a06707381bf2672e15b1db8058dc218bd8b2179b7e76f1ada53e"} Dec 05 08:31:30 crc kubenswrapper[4876]: I1205 08:31:30.721945 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qjjkm" event={"ID":"437c91ca-0472-481b-b58f-40269f4d2c18","Type":"ContainerStarted","Data":"a262f22b6ad78a9a9705406edada3708e789a0dead231a2c2d7050dd8c385596"} Dec 05 08:31:30 crc kubenswrapper[4876]: I1205 08:31:30.745843 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qjjkm" podStartSLOduration=2.066031675 podStartE2EDuration="4.745813652s" podCreationTimestamp="2025-12-05 08:31:26 +0000 UTC" firstStartedPulling="2025-12-05 08:31:27.516843498 +0000 UTC m=+952.005508120" lastFinishedPulling="2025-12-05 08:31:30.196625475 +0000 UTC m=+954.685290097" observedRunningTime="2025-12-05 08:31:30.737943476 +0000 UTC m=+955.226608138" watchObservedRunningTime="2025-12-05 08:31:30.745813652 +0000 UTC m=+955.234478314" Dec 05 08:31:32 crc kubenswrapper[4876]: I1205 08:31:32.204362 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-gvmp4" Dec 05 08:31:36 crc kubenswrapper[4876]: I1205 08:31:36.539870 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:36 crc kubenswrapper[4876]: I1205 08:31:36.541152 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:36 crc kubenswrapper[4876]: I1205 08:31:36.582278 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:36 crc kubenswrapper[4876]: I1205 08:31:36.789730 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-qjjkm" Dec 05 08:31:41 crc kubenswrapper[4876]: I1205 08:31:41.596074 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-fgbfx" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.450043 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4"] Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.451356 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.453455 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-5n662" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.453663 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-util\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.453721 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-bundle\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.453758 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfwz5\" (UniqueName: \"kubernetes.io/projected/01744036-ed48-4c66-b33a-a7fb0d95ab80-kube-api-access-cfwz5\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.458987 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4"] Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.554856 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-util\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.554898 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-bundle\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.554946 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfwz5\" (UniqueName: \"kubernetes.io/projected/01744036-ed48-4c66-b33a-a7fb0d95ab80-kube-api-access-cfwz5\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.555446 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-util\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.555592 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-bundle\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.577330 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfwz5\" (UniqueName: \"kubernetes.io/projected/01744036-ed48-4c66-b33a-a7fb0d95ab80-kube-api-access-cfwz5\") pod \"40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:42 crc kubenswrapper[4876]: I1205 08:31:42.767562 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:43 crc kubenswrapper[4876]: I1205 08:31:43.162978 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4"] Dec 05 08:31:43 crc kubenswrapper[4876]: I1205 08:31:43.802458 4876 generic.go:334] "Generic (PLEG): container finished" podID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerID="75848c9233209ceed782c10c9de62cfb1c62444ef4565415b4eb57f32bf09d89" exitCode=0 Dec 05 08:31:43 crc kubenswrapper[4876]: I1205 08:31:43.802638 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" event={"ID":"01744036-ed48-4c66-b33a-a7fb0d95ab80","Type":"ContainerDied","Data":"75848c9233209ceed782c10c9de62cfb1c62444ef4565415b4eb57f32bf09d89"} Dec 05 08:31:43 crc kubenswrapper[4876]: I1205 08:31:43.802776 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" event={"ID":"01744036-ed48-4c66-b33a-a7fb0d95ab80","Type":"ContainerStarted","Data":"05a945fb325bd9690f9537d9265da9478757e32cfcf230371127da0a683e5c23"} Dec 05 08:31:44 crc kubenswrapper[4876]: I1205 08:31:44.811093 4876 generic.go:334] "Generic (PLEG): container finished" podID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerID="114c3de3e1f79a3c06d4418add07ce13221bdf1f635658a170c5e6ec7992d108" exitCode=0 Dec 05 08:31:44 crc kubenswrapper[4876]: I1205 08:31:44.811140 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" event={"ID":"01744036-ed48-4c66-b33a-a7fb0d95ab80","Type":"ContainerDied","Data":"114c3de3e1f79a3c06d4418add07ce13221bdf1f635658a170c5e6ec7992d108"} Dec 05 08:31:45 crc kubenswrapper[4876]: I1205 08:31:45.820565 4876 generic.go:334] "Generic (PLEG): container finished" podID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerID="8c672963769510b50922369d87de28538520d647a9402bc44820ca1a9a255b4d" exitCode=0 Dec 05 08:31:45 crc kubenswrapper[4876]: I1205 08:31:45.820604 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" event={"ID":"01744036-ed48-4c66-b33a-a7fb0d95ab80","Type":"ContainerDied","Data":"8c672963769510b50922369d87de28538520d647a9402bc44820ca1a9a255b4d"} Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.056661 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.213658 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfwz5\" (UniqueName: \"kubernetes.io/projected/01744036-ed48-4c66-b33a-a7fb0d95ab80-kube-api-access-cfwz5\") pod \"01744036-ed48-4c66-b33a-a7fb0d95ab80\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.213795 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-bundle\") pod \"01744036-ed48-4c66-b33a-a7fb0d95ab80\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.214062 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-util\") pod \"01744036-ed48-4c66-b33a-a7fb0d95ab80\" (UID: \"01744036-ed48-4c66-b33a-a7fb0d95ab80\") " Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.216478 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-bundle" (OuterVolumeSpecName: "bundle") pod "01744036-ed48-4c66-b33a-a7fb0d95ab80" (UID: "01744036-ed48-4c66-b33a-a7fb0d95ab80"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.219483 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01744036-ed48-4c66-b33a-a7fb0d95ab80-kube-api-access-cfwz5" (OuterVolumeSpecName: "kube-api-access-cfwz5") pod "01744036-ed48-4c66-b33a-a7fb0d95ab80" (UID: "01744036-ed48-4c66-b33a-a7fb0d95ab80"). InnerVolumeSpecName "kube-api-access-cfwz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.244378 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-util" (OuterVolumeSpecName: "util") pod "01744036-ed48-4c66-b33a-a7fb0d95ab80" (UID: "01744036-ed48-4c66-b33a-a7fb0d95ab80"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.315925 4876 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-util\") on node \"crc\" DevicePath \"\"" Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.315972 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfwz5\" (UniqueName: \"kubernetes.io/projected/01744036-ed48-4c66-b33a-a7fb0d95ab80-kube-api-access-cfwz5\") on node \"crc\" DevicePath \"\"" Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.315985 4876 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/01744036-ed48-4c66-b33a-a7fb0d95ab80-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.836716 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.837590 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4" event={"ID":"01744036-ed48-4c66-b33a-a7fb0d95ab80","Type":"ContainerDied","Data":"05a945fb325bd9690f9537d9265da9478757e32cfcf230371127da0a683e5c23"} Dec 05 08:31:47 crc kubenswrapper[4876]: I1205 08:31:47.837632 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05a945fb325bd9690f9537d9265da9478757e32cfcf230371127da0a683e5c23" Dec 05 08:31:47 crc kubenswrapper[4876]: E1205 08:31:47.910939 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01744036_ed48_4c66_b33a_a7fb0d95ab80.slice/crio-05a945fb325bd9690f9537d9265da9478757e32cfcf230371127da0a683e5c23\": RecentStats: unable to find data in memory cache]" Dec 05 08:31:54 crc kubenswrapper[4876]: I1205 08:31:54.968260 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k"] Dec 05 08:31:54 crc kubenswrapper[4876]: E1205 08:31:54.969107 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerName="pull" Dec 05 08:31:54 crc kubenswrapper[4876]: I1205 08:31:54.969124 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerName="pull" Dec 05 08:31:54 crc kubenswrapper[4876]: E1205 08:31:54.969136 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerName="extract" Dec 05 08:31:54 crc kubenswrapper[4876]: I1205 08:31:54.969143 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerName="extract" Dec 05 08:31:54 crc kubenswrapper[4876]: E1205 08:31:54.969168 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerName="util" Dec 05 08:31:54 crc kubenswrapper[4876]: I1205 08:31:54.969176 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerName="util" Dec 05 08:31:54 crc kubenswrapper[4876]: I1205 08:31:54.969298 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="01744036-ed48-4c66-b33a-a7fb0d95ab80" containerName="extract" Dec 05 08:31:54 crc kubenswrapper[4876]: I1205 08:31:54.969763 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" Dec 05 08:31:54 crc kubenswrapper[4876]: I1205 08:31:54.971550 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-4j2v4" Dec 05 08:31:55 crc kubenswrapper[4876]: I1205 08:31:55.003227 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k"] Dec 05 08:31:55 crc kubenswrapper[4876]: I1205 08:31:55.114269 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qjw8\" (UniqueName: \"kubernetes.io/projected/1c603b75-2c70-4ac9-ba31-7f2f69785482-kube-api-access-7qjw8\") pod \"openstack-operator-controller-operator-cc6d987f6-whr8k\" (UID: \"1c603b75-2c70-4ac9-ba31-7f2f69785482\") " pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" Dec 05 08:31:55 crc kubenswrapper[4876]: I1205 08:31:55.215392 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qjw8\" (UniqueName: \"kubernetes.io/projected/1c603b75-2c70-4ac9-ba31-7f2f69785482-kube-api-access-7qjw8\") pod \"openstack-operator-controller-operator-cc6d987f6-whr8k\" (UID: \"1c603b75-2c70-4ac9-ba31-7f2f69785482\") " pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" Dec 05 08:31:55 crc kubenswrapper[4876]: I1205 08:31:55.238721 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qjw8\" (UniqueName: \"kubernetes.io/projected/1c603b75-2c70-4ac9-ba31-7f2f69785482-kube-api-access-7qjw8\") pod \"openstack-operator-controller-operator-cc6d987f6-whr8k\" (UID: \"1c603b75-2c70-4ac9-ba31-7f2f69785482\") " pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" Dec 05 08:31:55 crc kubenswrapper[4876]: I1205 08:31:55.289353 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" Dec 05 08:31:55 crc kubenswrapper[4876]: I1205 08:31:55.715956 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k"] Dec 05 08:31:55 crc kubenswrapper[4876]: I1205 08:31:55.882781 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" event={"ID":"1c603b75-2c70-4ac9-ba31-7f2f69785482","Type":"ContainerStarted","Data":"4453609078b07232bf9afef6c6c17700f981f7c8afbd793c9cd107dbf7b74762"} Dec 05 08:32:01 crc kubenswrapper[4876]: I1205 08:32:01.939506 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" event={"ID":"1c603b75-2c70-4ac9-ba31-7f2f69785482","Type":"ContainerStarted","Data":"bbd9dfb32accb70508c4fa7f719c68708dcca740d895471c9fd0a054a7be317c"} Dec 05 08:32:01 crc kubenswrapper[4876]: I1205 08:32:01.940139 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" Dec 05 08:32:01 crc kubenswrapper[4876]: I1205 08:32:01.975020 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" podStartSLOduration=2.926753158 podStartE2EDuration="7.975002988s" podCreationTimestamp="2025-12-05 08:31:54 +0000 UTC" firstStartedPulling="2025-12-05 08:31:55.724545546 +0000 UTC m=+980.213210178" lastFinishedPulling="2025-12-05 08:32:00.772795386 +0000 UTC m=+985.261460008" observedRunningTime="2025-12-05 08:32:01.972172133 +0000 UTC m=+986.460836765" watchObservedRunningTime="2025-12-05 08:32:01.975002988 +0000 UTC m=+986.463667610" Dec 05 08:32:05 crc kubenswrapper[4876]: I1205 08:32:05.292054 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-cc6d987f6-whr8k" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.846331 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.847717 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.849791 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-bhxrd" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.867748 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.868778 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.871735 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-825lc" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.872007 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.881528 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.882586 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.892294 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-5zn7r" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.896436 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.908953 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.915154 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.916145 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.943389 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-lbgdb" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.956954 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.963467 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.964616 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.968228 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-tdsfq" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.970593 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.971411 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.973695 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-chk8l" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.980323 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.982884 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndddk\" (UniqueName: \"kubernetes.io/projected/d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca-kube-api-access-ndddk\") pod \"designate-operator-controller-manager-78b4bc895b-jwn22\" (UID: \"d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.982953 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-774lt\" (UniqueName: \"kubernetes.io/projected/0036dfa3-4545-43f5-a6f6-c329485b678b-kube-api-access-774lt\") pod \"cinder-operator-controller-manager-859b6ccc6-d4l6m\" (UID: \"0036dfa3-4545-43f5-a6f6-c329485b678b\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.982995 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5qxs\" (UniqueName: \"kubernetes.io/projected/04e3edd8-f29a-48d1-84f4-81985747740f-kube-api-access-m5qxs\") pod \"barbican-operator-controller-manager-7d9dfd778-ppsdr\" (UID: \"04e3edd8-f29a-48d1-84f4-81985747740f\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.992579 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7"] Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.993668 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:24 crc kubenswrapper[4876]: I1205 08:32:24.996159 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-49vd6" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.006167 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.009065 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.013007 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.014122 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.018159 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.022488 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-cltt8" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.024366 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.037171 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.038163 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.041264 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-86jlc" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.049948 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.050983 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.052252 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-m5sxg" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.065953 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.074533 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.075569 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.080386 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-rgvkf" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.084141 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndddk\" (UniqueName: \"kubernetes.io/projected/d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca-kube-api-access-ndddk\") pod \"designate-operator-controller-manager-78b4bc895b-jwn22\" (UID: \"d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.084198 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-774lt\" (UniqueName: \"kubernetes.io/projected/0036dfa3-4545-43f5-a6f6-c329485b678b-kube-api-access-774lt\") pod \"cinder-operator-controller-manager-859b6ccc6-d4l6m\" (UID: \"0036dfa3-4545-43f5-a6f6-c329485b678b\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.084230 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59n45\" (UniqueName: \"kubernetes.io/projected/291c6269-dc8a-4193-a602-2a16202db817-kube-api-access-59n45\") pod \"heat-operator-controller-manager-5f64f6f8bb-q7r4r\" (UID: \"291c6269-dc8a-4193-a602-2a16202db817\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.084263 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wk7z\" (UniqueName: \"kubernetes.io/projected/679c172a-10ff-4a0e-8932-bcc1f78393d7-kube-api-access-8wk7z\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.084292 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5qxs\" (UniqueName: \"kubernetes.io/projected/04e3edd8-f29a-48d1-84f4-81985747740f-kube-api-access-m5qxs\") pod \"barbican-operator-controller-manager-7d9dfd778-ppsdr\" (UID: \"04e3edd8-f29a-48d1-84f4-81985747740f\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.084321 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4wbz\" (UniqueName: \"kubernetes.io/projected/1c4ffc06-424a-4c94-be0d-2ff27bf8d920-kube-api-access-r4wbz\") pod \"glance-operator-controller-manager-77987cd8cd-bv2mb\" (UID: \"1c4ffc06-424a-4c94-be0d-2ff27bf8d920\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.084359 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.084408 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd6wl\" (UniqueName: \"kubernetes.io/projected/0b8ca916-a8c5-4866-8333-ebbc0f861161-kube-api-access-qd6wl\") pod \"horizon-operator-controller-manager-68c6d99b8f-vjh2v\" (UID: \"0b8ca916-a8c5-4866-8333-ebbc0f861161\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.100970 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.101022 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.102300 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.103406 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.105333 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-r4t4j" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.111880 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.120937 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndddk\" (UniqueName: \"kubernetes.io/projected/d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca-kube-api-access-ndddk\") pod \"designate-operator-controller-manager-78b4bc895b-jwn22\" (UID: \"d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.121495 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-774lt\" (UniqueName: \"kubernetes.io/projected/0036dfa3-4545-43f5-a6f6-c329485b678b-kube-api-access-774lt\") pod \"cinder-operator-controller-manager-859b6ccc6-d4l6m\" (UID: \"0036dfa3-4545-43f5-a6f6-c329485b678b\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.122006 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5qxs\" (UniqueName: \"kubernetes.io/projected/04e3edd8-f29a-48d1-84f4-81985747740f-kube-api-access-m5qxs\") pod \"barbican-operator-controller-manager-7d9dfd778-ppsdr\" (UID: \"04e3edd8-f29a-48d1-84f4-81985747740f\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.129927 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.129960 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.130035 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.134646 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-svc6v"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.135486 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.135515 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-94sm6" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.140032 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-f4bwv" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.149945 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-svc6v"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.156732 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.157745 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.159799 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nlt6z" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.163931 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.185382 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.186461 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59n45\" (UniqueName: \"kubernetes.io/projected/291c6269-dc8a-4193-a602-2a16202db817-kube-api-access-59n45\") pod \"heat-operator-controller-manager-5f64f6f8bb-q7r4r\" (UID: \"291c6269-dc8a-4193-a602-2a16202db817\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.186564 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wk7z\" (UniqueName: \"kubernetes.io/projected/679c172a-10ff-4a0e-8932-bcc1f78393d7-kube-api-access-8wk7z\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.186730 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkjwj\" (UniqueName: \"kubernetes.io/projected/ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7-kube-api-access-rkjwj\") pod \"ironic-operator-controller-manager-6c548fd776-qgxmr\" (UID: \"ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.186830 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6w6k\" (UniqueName: \"kubernetes.io/projected/13b56ee6-b8fa-40bf-add8-1778f4acf30d-kube-api-access-x6w6k\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vp2x4\" (UID: \"13b56ee6-b8fa-40bf-add8-1778f4acf30d\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.186944 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4wbz\" (UniqueName: \"kubernetes.io/projected/1c4ffc06-424a-4c94-be0d-2ff27bf8d920-kube-api-access-r4wbz\") pod \"glance-operator-controller-manager-77987cd8cd-bv2mb\" (UID: \"1c4ffc06-424a-4c94-be0d-2ff27bf8d920\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.187000 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvg4w\" (UniqueName: \"kubernetes.io/projected/e21b6aa8-f703-42e0-8251-06444e6423fe-kube-api-access-zvg4w\") pod \"keystone-operator-controller-manager-7765d96ddf-4ckhg\" (UID: \"e21b6aa8-f703-42e0-8251-06444e6423fe\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.187053 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.187280 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.187335 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert podName:679c172a-10ff-4a0e-8932-bcc1f78393d7 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:25.687318539 +0000 UTC m=+1010.175983161 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert") pod "infra-operator-controller-manager-57548d458d-l9jr7" (UID: "679c172a-10ff-4a0e-8932-bcc1f78393d7") : secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.187512 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd6wl\" (UniqueName: \"kubernetes.io/projected/0b8ca916-a8c5-4866-8333-ebbc0f861161-kube-api-access-qd6wl\") pod \"horizon-operator-controller-manager-68c6d99b8f-vjh2v\" (UID: \"0b8ca916-a8c5-4866-8333-ebbc0f861161\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.187547 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlx9q\" (UniqueName: \"kubernetes.io/projected/9f436ce1-a9d9-411a-824a-85afc3ee2069-kube-api-access-xlx9q\") pod \"manila-operator-controller-manager-7c79b5df47-v2p7b\" (UID: \"9f436ce1-a9d9-411a-824a-85afc3ee2069\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.203187 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.206721 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.207912 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.220984 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.221466 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8llhg" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.221850 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.225459 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59n45\" (UniqueName: \"kubernetes.io/projected/291c6269-dc8a-4193-a602-2a16202db817-kube-api-access-59n45\") pod \"heat-operator-controller-manager-5f64f6f8bb-q7r4r\" (UID: \"291c6269-dc8a-4193-a602-2a16202db817\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.226519 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wptqj"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.227216 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wk7z\" (UniqueName: \"kubernetes.io/projected/679c172a-10ff-4a0e-8932-bcc1f78393d7-kube-api-access-8wk7z\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.228823 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.230488 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4wbz\" (UniqueName: \"kubernetes.io/projected/1c4ffc06-424a-4c94-be0d-2ff27bf8d920-kube-api-access-r4wbz\") pod \"glance-operator-controller-manager-77987cd8cd-bv2mb\" (UID: \"1c4ffc06-424a-4c94-be0d-2ff27bf8d920\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.233035 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-2b7zw" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.233816 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd6wl\" (UniqueName: \"kubernetes.io/projected/0b8ca916-a8c5-4866-8333-ebbc0f861161-kube-api-access-qd6wl\") pod \"horizon-operator-controller-manager-68c6d99b8f-vjh2v\" (UID: \"0b8ca916-a8c5-4866-8333-ebbc0f861161\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.252957 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wptqj"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.267629 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.286401 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.289287 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79rdn\" (UniqueName: \"kubernetes.io/projected/7b0b34b8-da13-4978-9a68-606135b29834-kube-api-access-79rdn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lxwt9\" (UID: \"7b0b34b8-da13-4978-9a68-606135b29834\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.289362 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhmn7\" (UniqueName: \"kubernetes.io/projected/68006b91-3398-4ff1-8c6e-ede02835c0ef-kube-api-access-qhmn7\") pod \"nova-operator-controller-manager-697bc559fc-2spcs\" (UID: \"68006b91-3398-4ff1-8c6e-ede02835c0ef\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.289446 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwhxg\" (UniqueName: \"kubernetes.io/projected/fb571bcf-872c-42f5-9a65-18c38291b062-kube-api-access-rwhxg\") pod \"octavia-operator-controller-manager-998648c74-svc6v\" (UID: \"fb571bcf-872c-42f5-9a65-18c38291b062\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.289498 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvg4w\" (UniqueName: \"kubernetes.io/projected/e21b6aa8-f703-42e0-8251-06444e6423fe-kube-api-access-zvg4w\") pod \"keystone-operator-controller-manager-7765d96ddf-4ckhg\" (UID: \"e21b6aa8-f703-42e0-8251-06444e6423fe\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.289636 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlx9q\" (UniqueName: \"kubernetes.io/projected/9f436ce1-a9d9-411a-824a-85afc3ee2069-kube-api-access-xlx9q\") pod \"manila-operator-controller-manager-7c79b5df47-v2p7b\" (UID: \"9f436ce1-a9d9-411a-824a-85afc3ee2069\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.289775 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn7ss\" (UniqueName: \"kubernetes.io/projected/217cf4c3-b49c-4bc6-a1cd-a12589264d11-kube-api-access-jn7ss\") pod \"ovn-operator-controller-manager-b6456fdb6-lfncq\" (UID: \"217cf4c3-b49c-4bc6-a1cd-a12589264d11\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.290234 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkjwj\" (UniqueName: \"kubernetes.io/projected/ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7-kube-api-access-rkjwj\") pod \"ironic-operator-controller-manager-6c548fd776-qgxmr\" (UID: \"ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.290282 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6w6k\" (UniqueName: \"kubernetes.io/projected/13b56ee6-b8fa-40bf-add8-1778f4acf30d-kube-api-access-x6w6k\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vp2x4\" (UID: \"13b56ee6-b8fa-40bf-add8-1778f4acf30d\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.298688 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.299804 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.303417 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-fdhpp" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.307528 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.311775 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6w6k\" (UniqueName: \"kubernetes.io/projected/13b56ee6-b8fa-40bf-add8-1778f4acf30d-kube-api-access-x6w6k\") pod \"mariadb-operator-controller-manager-56bbcc9d85-vp2x4\" (UID: \"13b56ee6-b8fa-40bf-add8-1778f4acf30d\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.313927 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvg4w\" (UniqueName: \"kubernetes.io/projected/e21b6aa8-f703-42e0-8251-06444e6423fe-kube-api-access-zvg4w\") pod \"keystone-operator-controller-manager-7765d96ddf-4ckhg\" (UID: \"e21b6aa8-f703-42e0-8251-06444e6423fe\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.314116 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlx9q\" (UniqueName: \"kubernetes.io/projected/9f436ce1-a9d9-411a-824a-85afc3ee2069-kube-api-access-xlx9q\") pod \"manila-operator-controller-manager-7c79b5df47-v2p7b\" (UID: \"9f436ce1-a9d9-411a-824a-85afc3ee2069\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.317435 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.324452 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkjwj\" (UniqueName: \"kubernetes.io/projected/ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7-kube-api-access-rkjwj\") pod \"ironic-operator-controller-manager-6c548fd776-qgxmr\" (UID: \"ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.330792 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.337946 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.344240 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.352039 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.352797 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-d9vqt" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.359796 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.369229 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.376974 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.378040 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.380437 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6xw59" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.387181 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.390575 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.393183 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kknd\" (UniqueName: \"kubernetes.io/projected/ca537166-1c8e-4616-ba9f-b97382f11d7f-kube-api-access-4kknd\") pod \"placement-operator-controller-manager-78f8948974-wptqj\" (UID: \"ca537166-1c8e-4616-ba9f-b97382f11d7f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.393221 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.393249 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn7ss\" (UniqueName: \"kubernetes.io/projected/217cf4c3-b49c-4bc6-a1cd-a12589264d11-kube-api-access-jn7ss\") pod \"ovn-operator-controller-manager-b6456fdb6-lfncq\" (UID: \"217cf4c3-b49c-4bc6-a1cd-a12589264d11\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.393286 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79rdn\" (UniqueName: \"kubernetes.io/projected/7b0b34b8-da13-4978-9a68-606135b29834-kube-api-access-79rdn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lxwt9\" (UID: \"7b0b34b8-da13-4978-9a68-606135b29834\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.393307 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwhxg\" (UniqueName: \"kubernetes.io/projected/fb571bcf-872c-42f5-9a65-18c38291b062-kube-api-access-rwhxg\") pod \"octavia-operator-controller-manager-998648c74-svc6v\" (UID: \"fb571bcf-872c-42f5-9a65-18c38291b062\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.393359 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhmn7\" (UniqueName: \"kubernetes.io/projected/68006b91-3398-4ff1-8c6e-ede02835c0ef-kube-api-access-qhmn7\") pod \"nova-operator-controller-manager-697bc559fc-2spcs\" (UID: \"68006b91-3398-4ff1-8c6e-ede02835c0ef\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.393441 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2rqs\" (UniqueName: \"kubernetes.io/projected/e3a38502-6fad-4383-81e8-591318a4d4d3-kube-api-access-d2rqs\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.393484 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcvsv\" (UniqueName: \"kubernetes.io/projected/861cead7-82be-495f-b144-7120ee08ae2d-kube-api-access-pcvsv\") pod \"swift-operator-controller-manager-5f8c65bbfc-qltmw\" (UID: \"861cead7-82be-495f-b144-7120ee08ae2d\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.424514 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.434225 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn7ss\" (UniqueName: \"kubernetes.io/projected/217cf4c3-b49c-4bc6-a1cd-a12589264d11-kube-api-access-jn7ss\") pod \"ovn-operator-controller-manager-b6456fdb6-lfncq\" (UID: \"217cf4c3-b49c-4bc6-a1cd-a12589264d11\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.434775 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwhxg\" (UniqueName: \"kubernetes.io/projected/fb571bcf-872c-42f5-9a65-18c38291b062-kube-api-access-rwhxg\") pod \"octavia-operator-controller-manager-998648c74-svc6v\" (UID: \"fb571bcf-872c-42f5-9a65-18c38291b062\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.442133 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhmn7\" (UniqueName: \"kubernetes.io/projected/68006b91-3398-4ff1-8c6e-ede02835c0ef-kube-api-access-qhmn7\") pod \"nova-operator-controller-manager-697bc559fc-2spcs\" (UID: \"68006b91-3398-4ff1-8c6e-ede02835c0ef\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.450462 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79rdn\" (UniqueName: \"kubernetes.io/projected/7b0b34b8-da13-4978-9a68-606135b29834-kube-api-access-79rdn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lxwt9\" (UID: \"7b0b34b8-da13-4978-9a68-606135b29834\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.458200 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.459497 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.461648 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-gvgnd" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.463247 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.470288 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.487157 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.495308 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kknd\" (UniqueName: \"kubernetes.io/projected/ca537166-1c8e-4616-ba9f-b97382f11d7f-kube-api-access-4kknd\") pod \"placement-operator-controller-manager-78f8948974-wptqj\" (UID: \"ca537166-1c8e-4616-ba9f-b97382f11d7f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.495348 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.495394 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4twdd\" (UniqueName: \"kubernetes.io/projected/74207150-aafb-409d-b5fb-7d56b9ac8f9c-kube-api-access-4twdd\") pod \"telemetry-operator-controller-manager-59c7b9d69d-t9vv2\" (UID: \"74207150-aafb-409d-b5fb-7d56b9ac8f9c\") " pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.495449 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2rqs\" (UniqueName: \"kubernetes.io/projected/e3a38502-6fad-4383-81e8-591318a4d4d3-kube-api-access-d2rqs\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.495494 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdjgl\" (UniqueName: \"kubernetes.io/projected/af99aad7-cbab-4990-8ed5-3a66c9f4b432-kube-api-access-cdjgl\") pod \"test-operator-controller-manager-5854674fcc-tn7mh\" (UID: \"af99aad7-cbab-4990-8ed5-3a66c9f4b432\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.495516 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcvsv\" (UniqueName: \"kubernetes.io/projected/861cead7-82be-495f-b144-7120ee08ae2d-kube-api-access-pcvsv\") pod \"swift-operator-controller-manager-5f8c65bbfc-qltmw\" (UID: \"861cead7-82be-495f-b144-7120ee08ae2d\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.495618 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.495695 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert podName:e3a38502-6fad-4383-81e8-591318a4d4d3 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:25.995675879 +0000 UTC m=+1010.484340501 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" (UID: "e3a38502-6fad-4383-81e8-591318a4d4d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.513332 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kknd\" (UniqueName: \"kubernetes.io/projected/ca537166-1c8e-4616-ba9f-b97382f11d7f-kube-api-access-4kknd\") pod \"placement-operator-controller-manager-78f8948974-wptqj\" (UID: \"ca537166-1c8e-4616-ba9f-b97382f11d7f\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.514770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2rqs\" (UniqueName: \"kubernetes.io/projected/e3a38502-6fad-4383-81e8-591318a4d4d3-kube-api-access-d2rqs\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.517332 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcvsv\" (UniqueName: \"kubernetes.io/projected/861cead7-82be-495f-b144-7120ee08ae2d-kube-api-access-pcvsv\") pod \"swift-operator-controller-manager-5f8c65bbfc-qltmw\" (UID: \"861cead7-82be-495f-b144-7120ee08ae2d\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.557094 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.575057 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.576011 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.578430 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.584333 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-lqrqm" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.584514 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.584614 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.592232 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.596398 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6b5d\" (UniqueName: \"kubernetes.io/projected/b2eec441-3198-440a-b061-47fe9bf9a340-kube-api-access-z6b5d\") pod \"watcher-operator-controller-manager-769dc69bc-hxc5v\" (UID: \"b2eec441-3198-440a-b061-47fe9bf9a340\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.596433 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4twdd\" (UniqueName: \"kubernetes.io/projected/74207150-aafb-409d-b5fb-7d56b9ac8f9c-kube-api-access-4twdd\") pod \"telemetry-operator-controller-manager-59c7b9d69d-t9vv2\" (UID: \"74207150-aafb-409d-b5fb-7d56b9ac8f9c\") " pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.596500 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdjgl\" (UniqueName: \"kubernetes.io/projected/af99aad7-cbab-4990-8ed5-3a66c9f4b432-kube-api-access-cdjgl\") pod \"test-operator-controller-manager-5854674fcc-tn7mh\" (UID: \"af99aad7-cbab-4990-8ed5-3a66c9f4b432\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.600203 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.614525 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdjgl\" (UniqueName: \"kubernetes.io/projected/af99aad7-cbab-4990-8ed5-3a66c9f4b432-kube-api-access-cdjgl\") pod \"test-operator-controller-manager-5854674fcc-tn7mh\" (UID: \"af99aad7-cbab-4990-8ed5-3a66c9f4b432\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.617244 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4twdd\" (UniqueName: \"kubernetes.io/projected/74207150-aafb-409d-b5fb-7d56b9ac8f9c-kube-api-access-4twdd\") pod \"telemetry-operator-controller-manager-59c7b9d69d-t9vv2\" (UID: \"74207150-aafb-409d-b5fb-7d56b9ac8f9c\") " pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.642531 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.654047 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.655080 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.660713 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-t4gm9" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.677567 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.689814 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.698327 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.698380 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jtcg\" (UniqueName: \"kubernetes.io/projected/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-kube-api-access-4jtcg\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.698420 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.698517 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.698568 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6b5d\" (UniqueName: \"kubernetes.io/projected/b2eec441-3198-440a-b061-47fe9bf9a340-kube-api-access-z6b5d\") pod \"watcher-operator-controller-manager-769dc69bc-hxc5v\" (UID: \"b2eec441-3198-440a-b061-47fe9bf9a340\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.704065 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.704138 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert podName:679c172a-10ff-4a0e-8932-bcc1f78393d7 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:26.704116229 +0000 UTC m=+1011.192780851 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert") pod "infra-operator-controller-manager-57548d458d-l9jr7" (UID: "679c172a-10ff-4a0e-8932-bcc1f78393d7") : secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.715616 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.739111 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6b5d\" (UniqueName: \"kubernetes.io/projected/b2eec441-3198-440a-b061-47fe9bf9a340-kube-api-access-z6b5d\") pod \"watcher-operator-controller-manager-769dc69bc-hxc5v\" (UID: \"b2eec441-3198-440a-b061-47fe9bf9a340\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.745222 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr"] Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.793048 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.802365 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.802419 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jtcg\" (UniqueName: \"kubernetes.io/projected/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-kube-api-access-4jtcg\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.802515 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.802559 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6dfm\" (UniqueName: \"kubernetes.io/projected/af4995ab-bc40-4446-bc2e-d421fe03423c-kube-api-access-l6dfm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9ptpl\" (UID: \"af4995ab-bc40-4446-bc2e-d421fe03423c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.802727 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.802796 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:26.302762216 +0000 UTC m=+1010.791426838 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "metrics-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.803280 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: E1205 08:32:25.803304 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:26.30329677 +0000 UTC m=+1010.791961392 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "webhook-server-cert" not found Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.828493 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.839802 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jtcg\" (UniqueName: \"kubernetes.io/projected/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-kube-api-access-4jtcg\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:25 crc kubenswrapper[4876]: I1205 08:32:25.905763 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6dfm\" (UniqueName: \"kubernetes.io/projected/af4995ab-bc40-4446-bc2e-d421fe03423c-kube-api-access-l6dfm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9ptpl\" (UID: \"af4995ab-bc40-4446-bc2e-d421fe03423c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:25.944587 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6dfm\" (UniqueName: \"kubernetes.io/projected/af4995ab-bc40-4446-bc2e-d421fe03423c-kube-api-access-l6dfm\") pod \"rabbitmq-cluster-operator-manager-668c99d594-9ptpl\" (UID: \"af4995ab-bc40-4446-bc2e-d421fe03423c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.008996 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.009153 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.009208 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert podName:e3a38502-6fad-4383-81e8-591318a4d4d3 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:27.009190582 +0000 UTC m=+1011.497855204 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" (UID: "e3a38502-6fad-4383-81e8-591318a4d4d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.040621 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.117863 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.124231 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" event={"ID":"04e3edd8-f29a-48d1-84f4-81985747740f","Type":"ContainerStarted","Data":"aedb3bddeb9e3430b113ae44d1f1f4130d46f9287d10cfa0c24042e2de96f5bf"} Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.129263 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" event={"ID":"0036dfa3-4545-43f5-a6f6-c329485b678b","Type":"ContainerStarted","Data":"92b62bbd1f284cd5373abf10f8dddd74ffe58cc9190ee33259356360ea6ec15b"} Dec 05 08:32:26 crc kubenswrapper[4876]: W1205 08:32:26.153519 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b1b3f4_28b5_44b8_a499_81e3fd64c8ca.slice/crio-bdeced5125a866045b9078fe8d3a22e37279879dde2cc4916b752063584d173d WatchSource:0}: Error finding container bdeced5125a866045b9078fe8d3a22e37279879dde2cc4916b752063584d173d: Status 404 returned error can't find the container with id bdeced5125a866045b9078fe8d3a22e37279879dde2cc4916b752063584d173d Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.314760 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.314856 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.315061 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.315122 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:27.315100168 +0000 UTC m=+1011.803764790 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "webhook-server-cert" not found Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.315167 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.315187 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:27.31518086 +0000 UTC m=+1011.803845482 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "metrics-server-cert" not found Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.373764 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.408125 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.436885 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.578230 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.606123 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.611812 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.665293 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.672835 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.679434 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.682288 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.689603 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9"] Dec 05 08:32:26 crc kubenswrapper[4876]: W1205 08:32:26.710373 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod291c6269_dc8a_4193_a602_2a16202db817.slice/crio-080bc6eece2d6161939999ea53d3d86fdaba22acf6a0f14d064e6ef4781550ef WatchSource:0}: Error finding container 080bc6eece2d6161939999ea53d3d86fdaba22acf6a0f14d064e6ef4781550ef: Status 404 returned error can't find the container with id 080bc6eece2d6161939999ea53d3d86fdaba22acf6a0f14d064e6ef4781550ef Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.720525 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.720687 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.720748 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert podName:679c172a-10ff-4a0e-8932-bcc1f78393d7 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:28.720729959 +0000 UTC m=+1013.209394581 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert") pod "infra-operator-controller-manager-57548d458d-l9jr7" (UID: "679c172a-10ff-4a0e-8932-bcc1f78393d7") : secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.740687 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-79rdn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-lxwt9_openstack-operators(7b0b34b8-da13-4978-9a68-606135b29834): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.742879 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-79rdn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-lxwt9_openstack-operators(7b0b34b8-da13-4978-9a68-606135b29834): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.744611 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" podUID="7b0b34b8-da13-4978-9a68-606135b29834" Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.818672 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.827604 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.845427 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh"] Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.850641 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wptqj"] Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.851813 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z6b5d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-hxc5v_openstack-operators(b2eec441-3198-440a-b061-47fe9bf9a340): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.854212 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z6b5d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-hxc5v_openstack-operators(b2eec441-3198-440a-b061-47fe9bf9a340): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.855362 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" podUID="b2eec441-3198-440a-b061-47fe9bf9a340" Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.857913 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw"] Dec 05 08:32:26 crc kubenswrapper[4876]: W1205 08:32:26.860262 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf4995ab_bc40_4446_bc2e_d421fe03423c.slice/crio-f173b27e109d4aef1fef7e852ce17da047c1941fb25f99131e46c1eed1ab15cd WatchSource:0}: Error finding container f173b27e109d4aef1fef7e852ce17da047c1941fb25f99131e46c1eed1ab15cd: Status 404 returned error can't find the container with id f173b27e109d4aef1fef7e852ce17da047c1941fb25f99131e46c1eed1ab15cd Dec 05 08:32:26 crc kubenswrapper[4876]: W1205 08:32:26.860506 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf99aad7_cbab_4990_8ed5_3a66c9f4b432.slice/crio-c5d2335148a1a070ebb6ce96f4684ee7048bdfb95986423fe754d0cc64f1ac9a WatchSource:0}: Error finding container c5d2335148a1a070ebb6ce96f4684ee7048bdfb95986423fe754d0cc64f1ac9a: Status 404 returned error can't find the container with id c5d2335148a1a070ebb6ce96f4684ee7048bdfb95986423fe754d0cc64f1ac9a Dec 05 08:32:26 crc kubenswrapper[4876]: I1205 08:32:26.862315 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-svc6v"] Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.866268 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4kknd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-wptqj_openstack-operators(ca537166-1c8e-4616-ba9f-b97382f11d7f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.868519 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cdjgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-tn7mh_openstack-operators(af99aad7-cbab-4990-8ed5-3a66c9f4b432): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.869130 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4kknd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-wptqj_openstack-operators(ca537166-1c8e-4616-ba9f-b97382f11d7f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.869513 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rwhxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-svc6v_openstack-operators(fb571bcf-872c-42f5-9a65-18c38291b062): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.870509 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cdjgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-tn7mh_openstack-operators(af99aad7-cbab-4990-8ed5-3a66c9f4b432): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.870630 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" podUID="ca537166-1c8e-4616-ba9f-b97382f11d7f" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.871643 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" podUID="af99aad7-cbab-4990-8ed5-3a66c9f4b432" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.871673 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rwhxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-svc6v_openstack-operators(fb571bcf-872c-42f5-9a65-18c38291b062): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.873203 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" podUID="fb571bcf-872c-42f5-9a65-18c38291b062" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.873731 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l6dfm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-9ptpl_openstack-operators(af4995ab-bc40-4446-bc2e-d421fe03423c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.874834 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" podUID="af4995ab-bc40-4446-bc2e-d421fe03423c" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.877270 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pcvsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-qltmw_openstack-operators(861cead7-82be-495f-b144-7120ee08ae2d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.878933 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pcvsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-qltmw_openstack-operators(861cead7-82be-495f-b144-7120ee08ae2d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 08:32:26 crc kubenswrapper[4876]: E1205 08:32:26.880402 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" podUID="861cead7-82be-495f-b144-7120ee08ae2d" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.024131 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.024596 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.024679 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert podName:e3a38502-6fad-4383-81e8-591318a4d4d3 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:29.024634391 +0000 UTC m=+1013.513299013 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" (UID: "e3a38502-6fad-4383-81e8-591318a4d4d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.136786 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" event={"ID":"b2eec441-3198-440a-b061-47fe9bf9a340","Type":"ContainerStarted","Data":"d50872a890351537267b93f954c318b2f641c9dbfa57af09d803ffbcc2f1236c"} Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.140302 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" podUID="b2eec441-3198-440a-b061-47fe9bf9a340" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.143415 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" event={"ID":"9f436ce1-a9d9-411a-824a-85afc3ee2069","Type":"ContainerStarted","Data":"1affafa816953ec72be4bb77f2fff26ceaf262bc95a45e990c6220efde55f35c"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.147245 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" event={"ID":"217cf4c3-b49c-4bc6-a1cd-a12589264d11","Type":"ContainerStarted","Data":"d148fa2707e2ef2c4aa832db14ff3bb8457090e35fcf6baf021d16c657389859"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.159498 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" event={"ID":"291c6269-dc8a-4193-a602-2a16202db817","Type":"ContainerStarted","Data":"080bc6eece2d6161939999ea53d3d86fdaba22acf6a0f14d064e6ef4781550ef"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.173057 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" event={"ID":"861cead7-82be-495f-b144-7120ee08ae2d","Type":"ContainerStarted","Data":"c272a11a920753c58199727875475e6be9f364ad5fc6728994737ca5271dd0ea"} Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.175095 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" podUID="861cead7-82be-495f-b144-7120ee08ae2d" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.178994 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" event={"ID":"13b56ee6-b8fa-40bf-add8-1778f4acf30d","Type":"ContainerStarted","Data":"fc12252b7ece3695f218ff641fcd4235a3340177b10f16a29fb474dbdfe6c134"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.182736 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" event={"ID":"fb571bcf-872c-42f5-9a65-18c38291b062","Type":"ContainerStarted","Data":"0ec95d4ede5e2b20767c86cb5c7b6069a4550128a44298882fed59fb15371b7a"} Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.187596 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" podUID="fb571bcf-872c-42f5-9a65-18c38291b062" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.191200 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" event={"ID":"74207150-aafb-409d-b5fb-7d56b9ac8f9c","Type":"ContainerStarted","Data":"1301ce0a636c82c0f3452eea8537b1b7be4c460693e91aeb3c2a9cf607a160e8"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.202570 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" event={"ID":"0b8ca916-a8c5-4866-8333-ebbc0f861161","Type":"ContainerStarted","Data":"31c08384349f3ccee4aa00d200b46de813669d9303fc41f7e2da5172dc478dc9"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.211461 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" event={"ID":"ca537166-1c8e-4616-ba9f-b97382f11d7f","Type":"ContainerStarted","Data":"f80f0b0d219f7c39c0e83dcc6408b68f32ae45aa0502821dacb06c9fd3f03d07"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.213759 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" event={"ID":"68006b91-3398-4ff1-8c6e-ede02835c0ef","Type":"ContainerStarted","Data":"d3eb9348dabc44c3dbb493775c3a1c22c4e2a878727c956019e186c752c97309"} Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.226310 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" podUID="ca537166-1c8e-4616-ba9f-b97382f11d7f" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.259233 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" event={"ID":"af99aad7-cbab-4990-8ed5-3a66c9f4b432","Type":"ContainerStarted","Data":"c5d2335148a1a070ebb6ce96f4684ee7048bdfb95986423fe754d0cc64f1ac9a"} Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.261411 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" podUID="af99aad7-cbab-4990-8ed5-3a66c9f4b432" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.261864 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" event={"ID":"1c4ffc06-424a-4c94-be0d-2ff27bf8d920","Type":"ContainerStarted","Data":"790cf078e52440cd591e434c857f99c09276dbdd589ba19f432315ced7998e4c"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.271168 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" event={"ID":"af4995ab-bc40-4446-bc2e-d421fe03423c","Type":"ContainerStarted","Data":"f173b27e109d4aef1fef7e852ce17da047c1941fb25f99131e46c1eed1ab15cd"} Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.272694 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" podUID="af4995ab-bc40-4446-bc2e-d421fe03423c" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.280739 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" event={"ID":"7b0b34b8-da13-4978-9a68-606135b29834","Type":"ContainerStarted","Data":"1c7a2dd56053511ca59650195295c5e94db9d59c1ec32f1d76fe06107e8c42f4"} Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.297145 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" podUID="7b0b34b8-da13-4978-9a68-606135b29834" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.302236 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" event={"ID":"e21b6aa8-f703-42e0-8251-06444e6423fe","Type":"ContainerStarted","Data":"2a4d62c7672148be8b440090c435166020663fadef23292357e4f47d7affd765"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.322073 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" event={"ID":"ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7","Type":"ContainerStarted","Data":"585fac478f8833b395ee5c5867cb6670e36f6514e7b72588635a5e6f8b56815c"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.356329 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" event={"ID":"d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca","Type":"ContainerStarted","Data":"bdeced5125a866045b9078fe8d3a22e37279879dde2cc4916b752063584d173d"} Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.360203 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:27 crc kubenswrapper[4876]: I1205 08:32:27.360301 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.360940 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.360983 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:29.360969533 +0000 UTC m=+1013.849634145 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "webhook-server-cert" not found Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.361411 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 08:32:27 crc kubenswrapper[4876]: E1205 08:32:27.361439 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:29.361431846 +0000 UTC m=+1013.850096468 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "metrics-server-cert" not found Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.369403 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" podUID="af4995ab-bc40-4446-bc2e-d421fe03423c" Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.369765 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" podUID="af99aad7-cbab-4990-8ed5-3a66c9f4b432" Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.369952 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" podUID="7b0b34b8-da13-4978-9a68-606135b29834" Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.370028 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" podUID="fb571bcf-872c-42f5-9a65-18c38291b062" Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.370291 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" podUID="ca537166-1c8e-4616-ba9f-b97382f11d7f" Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.370392 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" podUID="861cead7-82be-495f-b144-7120ee08ae2d" Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.375699 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" podUID="b2eec441-3198-440a-b061-47fe9bf9a340" Dec 05 08:32:28 crc kubenswrapper[4876]: I1205 08:32:28.785933 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.786296 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:28 crc kubenswrapper[4876]: E1205 08:32:28.786344 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert podName:679c172a-10ff-4a0e-8932-bcc1f78393d7 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:32.786329336 +0000 UTC m=+1017.274993958 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert") pod "infra-operator-controller-manager-57548d458d-l9jr7" (UID: "679c172a-10ff-4a0e-8932-bcc1f78393d7") : secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:29 crc kubenswrapper[4876]: I1205 08:32:29.090521 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:29 crc kubenswrapper[4876]: E1205 08:32:29.090689 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:29 crc kubenswrapper[4876]: E1205 08:32:29.090756 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert podName:e3a38502-6fad-4383-81e8-591318a4d4d3 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:33.090737971 +0000 UTC m=+1017.579402593 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" (UID: "e3a38502-6fad-4383-81e8-591318a4d4d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:29 crc kubenswrapper[4876]: I1205 08:32:29.397287 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:29 crc kubenswrapper[4876]: I1205 08:32:29.397394 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:29 crc kubenswrapper[4876]: E1205 08:32:29.397535 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 08:32:29 crc kubenswrapper[4876]: E1205 08:32:29.397582 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:33.397567311 +0000 UTC m=+1017.886231933 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "webhook-server-cert" not found Dec 05 08:32:29 crc kubenswrapper[4876]: E1205 08:32:29.397621 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 08:32:29 crc kubenswrapper[4876]: E1205 08:32:29.397638 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:33.397633043 +0000 UTC m=+1017.886297665 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "metrics-server-cert" not found Dec 05 08:32:32 crc kubenswrapper[4876]: I1205 08:32:32.850759 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:32 crc kubenswrapper[4876]: E1205 08:32:32.851277 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:32 crc kubenswrapper[4876]: E1205 08:32:32.851330 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert podName:679c172a-10ff-4a0e-8932-bcc1f78393d7 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:40.851312834 +0000 UTC m=+1025.339977456 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert") pod "infra-operator-controller-manager-57548d458d-l9jr7" (UID: "679c172a-10ff-4a0e-8932-bcc1f78393d7") : secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:33 crc kubenswrapper[4876]: I1205 08:32:33.154866 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:33 crc kubenswrapper[4876]: E1205 08:32:33.155025 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:33 crc kubenswrapper[4876]: E1205 08:32:33.155070 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert podName:e3a38502-6fad-4383-81e8-591318a4d4d3 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:41.155056451 +0000 UTC m=+1025.643721073 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" (UID: "e3a38502-6fad-4383-81e8-591318a4d4d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:33 crc kubenswrapper[4876]: I1205 08:32:33.460605 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:33 crc kubenswrapper[4876]: I1205 08:32:33.460807 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:33 crc kubenswrapper[4876]: E1205 08:32:33.460844 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 08:32:33 crc kubenswrapper[4876]: E1205 08:32:33.460971 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:41.460945926 +0000 UTC m=+1025.949610588 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "metrics-server-cert" not found Dec 05 08:32:33 crc kubenswrapper[4876]: E1205 08:32:33.461034 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 08:32:33 crc kubenswrapper[4876]: E1205 08:32:33.461120 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:41.46110126 +0000 UTC m=+1025.949765962 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "webhook-server-cert" not found Dec 05 08:32:40 crc kubenswrapper[4876]: I1205 08:32:40.904409 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:40 crc kubenswrapper[4876]: E1205 08:32:40.904589 4876 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:40 crc kubenswrapper[4876]: E1205 08:32:40.905128 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert podName:679c172a-10ff-4a0e-8932-bcc1f78393d7 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:56.905104468 +0000 UTC m=+1041.393769090 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert") pod "infra-operator-controller-manager-57548d458d-l9jr7" (UID: "679c172a-10ff-4a0e-8932-bcc1f78393d7") : secret "infra-operator-webhook-server-cert" not found Dec 05 08:32:41 crc kubenswrapper[4876]: I1205 08:32:41.210202 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.210385 4876 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.210458 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert podName:e3a38502-6fad-4383-81e8-591318a4d4d3 nodeName:}" failed. No retries permitted until 2025-12-05 08:32:57.210439738 +0000 UTC m=+1041.699104360 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" (UID: "e3a38502-6fad-4383-81e8-591318a4d4d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.249679 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.249868 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rkjwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-qgxmr_openstack-operators(ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:32:41 crc kubenswrapper[4876]: I1205 08:32:41.513708 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:41 crc kubenswrapper[4876]: I1205 08:32:41.513839 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.514006 4876 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.514097 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:57.514079393 +0000 UTC m=+1042.002744015 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "webhook-server-cert" not found Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.514332 4876 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.514395 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs podName:48cca8bf-c0e1-4957-885a-4cf42c1c33fe nodeName:}" failed. No retries permitted until 2025-12-05 08:32:57.514383892 +0000 UTC m=+1042.003048514 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs") pod "openstack-operator-controller-manager-8579f747db-9t4h4" (UID: "48cca8bf-c0e1-4957-885a-4cf42c1c33fe") : secret "metrics-server-cert" not found Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.782219 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 05 08:32:41 crc kubenswrapper[4876]: E1205 08:32:41.782784 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x6w6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-vp2x4_openstack-operators(13b56ee6-b8fa-40bf-add8-1778f4acf30d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:32:42 crc kubenswrapper[4876]: E1205 08:32:42.447327 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 05 08:32:42 crc kubenswrapper[4876]: E1205 08:32:42.447531 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qd6wl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-vjh2v_openstack-operators(0b8ca916-a8c5-4866-8333-ebbc0f861161): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:32:43 crc kubenswrapper[4876]: E1205 08:32:43.136634 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 05 08:32:43 crc kubenswrapper[4876]: E1205 08:32:43.136831 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zvg4w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-4ckhg_openstack-operators(e21b6aa8-f703-42e0-8251-06444e6423fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:32:43 crc kubenswrapper[4876]: E1205 08:32:43.732160 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 05 08:32:43 crc kubenswrapper[4876]: E1205 08:32:43.732656 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qhmn7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-2spcs_openstack-operators(68006b91-3398-4ff1-8c6e-ede02835c0ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:32:50 crc kubenswrapper[4876]: I1205 08:32:50.530113 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" event={"ID":"1c4ffc06-424a-4c94-be0d-2ff27bf8d920","Type":"ContainerStarted","Data":"56fee8f3feab3ccb4b6c9c000ba7b766a81ebc5ef413abf4ab69e9c741b04251"} Dec 05 08:32:50 crc kubenswrapper[4876]: I1205 08:32:50.531931 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" event={"ID":"9f436ce1-a9d9-411a-824a-85afc3ee2069","Type":"ContainerStarted","Data":"0584e8a111f15cf63c5247e52f68e21cf2198a0026e02a276ed7ed64f339280e"} Dec 05 08:32:50 crc kubenswrapper[4876]: I1205 08:32:50.533399 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" event={"ID":"0036dfa3-4545-43f5-a6f6-c329485b678b","Type":"ContainerStarted","Data":"710994272b738e115a03e2f3fc35e61e99dcd652366383d324c6a92e15228676"} Dec 05 08:32:51 crc kubenswrapper[4876]: I1205 08:32:51.543594 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" event={"ID":"d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca","Type":"ContainerStarted","Data":"ee67deedbef18eeac92b815234e8be1f5695ad7b157e42230e7b93658aa04377"} Dec 05 08:32:51 crc kubenswrapper[4876]: I1205 08:32:51.555950 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" event={"ID":"74207150-aafb-409d-b5fb-7d56b9ac8f9c","Type":"ContainerStarted","Data":"8c343e07d655b1d7d0a3f3e3b7cb202759da59a865e6f106feccbce0e61784b4"} Dec 05 08:32:51 crc kubenswrapper[4876]: I1205 08:32:51.582252 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" event={"ID":"217cf4c3-b49c-4bc6-a1cd-a12589264d11","Type":"ContainerStarted","Data":"66f872aea4a3e247cc796930896106c1e1c438f99de3b778bf95f65f0144e932"} Dec 05 08:32:51 crc kubenswrapper[4876]: I1205 08:32:51.772498 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" event={"ID":"291c6269-dc8a-4193-a602-2a16202db817","Type":"ContainerStarted","Data":"1574fc76ca2a58c9ec54c39710347c014eac40e354bee4cb0c1a124de6296da5"} Dec 05 08:32:51 crc kubenswrapper[4876]: I1205 08:32:51.789580 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" event={"ID":"04e3edd8-f29a-48d1-84f4-81985747740f","Type":"ContainerStarted","Data":"4ca1f03aecbe383e88396ecf500cc95de1ea9fa903bfa5b6c06def8cf59afb6f"} Dec 05 08:32:52 crc kubenswrapper[4876]: I1205 08:32:52.839656 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" event={"ID":"861cead7-82be-495f-b144-7120ee08ae2d","Type":"ContainerStarted","Data":"b79b68f6ded41527b32ae4837f7b68ba8454e4ebc3dc9549b3f934956b5af5b3"} Dec 05 08:32:52 crc kubenswrapper[4876]: I1205 08:32:52.842238 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" event={"ID":"ca537166-1c8e-4616-ba9f-b97382f11d7f","Type":"ContainerStarted","Data":"79af501886bae5c65a7225fba6ceb1be0af47ca749725798729b00543d55756e"} Dec 05 08:32:52 crc kubenswrapper[4876]: I1205 08:32:52.844079 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" event={"ID":"7b0b34b8-da13-4978-9a68-606135b29834","Type":"ContainerStarted","Data":"1f7344947abd4616ab2d5f2d793e18ad3c8615f711b8c5d117634779e9a59f5b"} Dec 05 08:32:52 crc kubenswrapper[4876]: I1205 08:32:52.846043 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" event={"ID":"af99aad7-cbab-4990-8ed5-3a66c9f4b432","Type":"ContainerStarted","Data":"c0449eb2dfe49271c3f5458c30adcfe359753047236873c3f28c7b8fc898a274"} Dec 05 08:32:52 crc kubenswrapper[4876]: I1205 08:32:52.847545 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" event={"ID":"fb571bcf-872c-42f5-9a65-18c38291b062","Type":"ContainerStarted","Data":"a5e4d0bb2fdbf972ce76a57ed2654743f08681ea429c740f86a29f548a683e05"} Dec 05 08:32:54 crc kubenswrapper[4876]: E1205 08:32:54.154143 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" podUID="13b56ee6-b8fa-40bf-add8-1778f4acf30d" Dec 05 08:32:54 crc kubenswrapper[4876]: E1205 08:32:54.400187 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" podUID="e21b6aa8-f703-42e0-8251-06444e6423fe" Dec 05 08:32:54 crc kubenswrapper[4876]: E1205 08:32:54.689165 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" podUID="0b8ca916-a8c5-4866-8333-ebbc0f861161" Dec 05 08:32:54 crc kubenswrapper[4876]: E1205 08:32:54.732289 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" podUID="68006b91-3398-4ff1-8c6e-ede02835c0ef" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.868331 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" event={"ID":"04e3edd8-f29a-48d1-84f4-81985747740f","Type":"ContainerStarted","Data":"22b21518f568330a175c7c9731ee6fa1141315cb6dd8f22eccfe6170bfe36e60"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.868685 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.871967 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" event={"ID":"0036dfa3-4545-43f5-a6f6-c329485b678b","Type":"ContainerStarted","Data":"5473c9bd7d677f5cdec52d277a9802d13b44fc94df38dd35896babf065cbc6e8"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.872624 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.877569 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" event={"ID":"13b56ee6-b8fa-40bf-add8-1778f4acf30d","Type":"ContainerStarted","Data":"972e98dc25cf02bda3edba6b155165a5643e7c9a1f169265deb323168c2f7412"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.879720 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" event={"ID":"af99aad7-cbab-4990-8ed5-3a66c9f4b432","Type":"ContainerStarted","Data":"2da621c850e9173e508e9877d3fcf2358975d642521fd4fe0ce625b4dab7eaf5"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.880323 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.885172 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.890884 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" event={"ID":"1c4ffc06-424a-4c94-be0d-2ff27bf8d920","Type":"ContainerStarted","Data":"76b157652a2affc9ca16a63c6e183e412474720fcbbd55c70af14d2f63b56ca7"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.892494 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.897030 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.898312 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" podStartSLOduration=2.389379871 podStartE2EDuration="30.898299657s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:25.754487375 +0000 UTC m=+1010.243151997" lastFinishedPulling="2025-12-05 08:32:54.263407161 +0000 UTC m=+1038.752071783" observedRunningTime="2025-12-05 08:32:54.893572583 +0000 UTC m=+1039.382237205" watchObservedRunningTime="2025-12-05 08:32:54.898299657 +0000 UTC m=+1039.386964279" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.916596 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" event={"ID":"e21b6aa8-f703-42e0-8251-06444e6423fe","Type":"ContainerStarted","Data":"2ac82b99b403dee1f1c4f2b4e74901cb08a211db183a9ef7a304dc3b61e40252"} Dec 05 08:32:54 crc kubenswrapper[4876]: E1205 08:32:54.930716 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" podUID="ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.931772 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" event={"ID":"af4995ab-bc40-4446-bc2e-d421fe03423c","Type":"ContainerStarted","Data":"f5b48a8d32074f704a67f12896eefa142f433607fd7fcf1bc52a4ed254f0d6bb"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.936220 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-d4l6m" podStartSLOduration=2.54066096 podStartE2EDuration="30.93620176s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:25.763552586 +0000 UTC m=+1010.252217208" lastFinishedPulling="2025-12-05 08:32:54.159093386 +0000 UTC m=+1038.647758008" observedRunningTime="2025-12-05 08:32:54.933381356 +0000 UTC m=+1039.422045978" watchObservedRunningTime="2025-12-05 08:32:54.93620176 +0000 UTC m=+1039.424866382" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.951635 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" event={"ID":"74207150-aafb-409d-b5fb-7d56b9ac8f9c","Type":"ContainerStarted","Data":"32c99e9f36d5bb141bcfe602d930740fd2ac6d6b39682632f40bf06d50959b2e"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.952369 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.953991 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" event={"ID":"68006b91-3398-4ff1-8c6e-ede02835c0ef","Type":"ContainerStarted","Data":"02edcfde2c55437c8132935f50dfdb2411bc7428ce70844d21dcb4f2babfeb08"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.978391 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" event={"ID":"fb571bcf-872c-42f5-9a65-18c38291b062","Type":"ContainerStarted","Data":"40a811774ca4b318cd8681ec5e93f736a1bf4909ca58dfcbecd63fc355e3310f"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.979024 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.981535 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" event={"ID":"0b8ca916-a8c5-4866-8333-ebbc0f861161","Type":"ContainerStarted","Data":"5d9ef4f5edff6740a92aa5012f1e842b0f981ead4b88006a552f6574418829ad"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.993288 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" event={"ID":"291c6269-dc8a-4193-a602-2a16202db817","Type":"ContainerStarted","Data":"6e9a0b7d66b52338680403f7784b27208615f079103555baa35bfb586f9705c6"} Dec 05 08:32:54 crc kubenswrapper[4876]: I1205 08:32:54.994109 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.021002 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" podStartSLOduration=2.850145625 podStartE2EDuration="30.020978293s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.868407626 +0000 UTC m=+1011.357072248" lastFinishedPulling="2025-12-05 08:32:54.039240294 +0000 UTC m=+1038.527904916" observedRunningTime="2025-12-05 08:32:54.979540456 +0000 UTC m=+1039.468205078" watchObservedRunningTime="2025-12-05 08:32:55.020978293 +0000 UTC m=+1039.509642915" Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.022170 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-bv2mb" podStartSLOduration=3.5620769450000003 podStartE2EDuration="31.022164404s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.714099133 +0000 UTC m=+1011.202763755" lastFinishedPulling="2025-12-05 08:32:54.174186592 +0000 UTC m=+1038.662851214" observedRunningTime="2025-12-05 08:32:55.01097474 +0000 UTC m=+1039.499639372" watchObservedRunningTime="2025-12-05 08:32:55.022164404 +0000 UTC m=+1039.510829016" Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.046147 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" event={"ID":"b2eec441-3198-440a-b061-47fe9bf9a340","Type":"ContainerStarted","Data":"67722935b019ca5f3566b4d9c96576c5f345b512e86ee987ffb3658fafae4aa9"} Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.046193 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" event={"ID":"b2eec441-3198-440a-b061-47fe9bf9a340","Type":"ContainerStarted","Data":"e7d0c16783653ecd92743f5cf339b329171d1b670eaebb833039d88c4e8cd14f"} Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.046802 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.086025 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" podStartSLOduration=2.941541122 podStartE2EDuration="30.086002188s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.869412623 +0000 UTC m=+1011.358077245" lastFinishedPulling="2025-12-05 08:32:54.013873689 +0000 UTC m=+1038.502538311" observedRunningTime="2025-12-05 08:32:55.077507935 +0000 UTC m=+1039.566172567" watchObservedRunningTime="2025-12-05 08:32:55.086002188 +0000 UTC m=+1039.574666830" Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.199451 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" podStartSLOduration=3.865982831 podStartE2EDuration="31.199435771s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.713338382 +0000 UTC m=+1011.202003004" lastFinishedPulling="2025-12-05 08:32:54.046791312 +0000 UTC m=+1038.535455944" observedRunningTime="2025-12-05 08:32:55.198063515 +0000 UTC m=+1039.686728137" watchObservedRunningTime="2025-12-05 08:32:55.199435771 +0000 UTC m=+1039.688100393" Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.258042 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" podStartSLOduration=2.946717104 podStartE2EDuration="30.258023617s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.732129091 +0000 UTC m=+1011.220793713" lastFinishedPulling="2025-12-05 08:32:54.043435604 +0000 UTC m=+1038.532100226" observedRunningTime="2025-12-05 08:32:55.243143987 +0000 UTC m=+1039.731808609" watchObservedRunningTime="2025-12-05 08:32:55.258023617 +0000 UTC m=+1039.746688239" Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.334957 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-9ptpl" podStartSLOduration=7.407422883 podStartE2EDuration="30.334942454s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.873664976 +0000 UTC m=+1011.362329598" lastFinishedPulling="2025-12-05 08:32:49.801184547 +0000 UTC m=+1034.289849169" observedRunningTime="2025-12-05 08:32:55.330097857 +0000 UTC m=+1039.818762489" watchObservedRunningTime="2025-12-05 08:32:55.334942454 +0000 UTC m=+1039.823607076" Dec 05 08:32:55 crc kubenswrapper[4876]: I1205 08:32:55.354486 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" podStartSLOduration=7.830752666 podStartE2EDuration="30.354467256s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.851704923 +0000 UTC m=+1011.340369545" lastFinishedPulling="2025-12-05 08:32:49.375419503 +0000 UTC m=+1033.864084135" observedRunningTime="2025-12-05 08:32:55.350172353 +0000 UTC m=+1039.838836975" watchObservedRunningTime="2025-12-05 08:32:55.354467256 +0000 UTC m=+1039.843131878" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.057139 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" event={"ID":"861cead7-82be-495f-b144-7120ee08ae2d","Type":"ContainerStarted","Data":"bb243f8f55306ba61a121f006a849e46886e99906a3462788de6db9d925d20ad"} Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.057340 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.059832 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" event={"ID":"ca537166-1c8e-4616-ba9f-b97382f11d7f","Type":"ContainerStarted","Data":"fd8a0e204701220d3c554464dc95c8e212f565e591057391296d5f4afea673b9"} Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.059883 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.061668 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" event={"ID":"7b0b34b8-da13-4978-9a68-606135b29834","Type":"ContainerStarted","Data":"8b3547774a568317ed16904c73512efbdedf4892131aa90dcbd31891eacc38a1"} Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.062511 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.063350 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" event={"ID":"ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7","Type":"ContainerStarted","Data":"12d422977d1dbb897ad173ffdf5bafc5d0e7a8eaa95a80a2c821aec6d95bfffe"} Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.065387 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" event={"ID":"d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca","Type":"ContainerStarted","Data":"da6b7ad53c2fd1c8f5ab0ba8bcfea7883d598d620e46d631e4ee7a181d817bf5"} Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.065706 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.068099 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.068760 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" event={"ID":"9f436ce1-a9d9-411a-824a-85afc3ee2069","Type":"ContainerStarted","Data":"03dd0cd8f55f5b5daf53b41a7e52c83584eb9e4ec402e75332b2ba87867a49fc"} Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.070633 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.078293 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.079850 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" event={"ID":"217cf4c3-b49c-4bc6-a1cd-a12589264d11","Type":"ContainerStarted","Data":"fc2fc0518114eb505c96403545eedea82e78e38f608e73b18260782bf583d2ec"} Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.081660 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.082609 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ppsdr" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.082730 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-59c7b9d69d-t9vv2" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.083012 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-q7r4r" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.083586 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.104324 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" podStartSLOduration=3.805790442 podStartE2EDuration="31.104303385s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.877187659 +0000 UTC m=+1011.365852281" lastFinishedPulling="2025-12-05 08:32:54.175700602 +0000 UTC m=+1038.664365224" observedRunningTime="2025-12-05 08:32:56.096167481 +0000 UTC m=+1040.584832113" watchObservedRunningTime="2025-12-05 08:32:56.104303385 +0000 UTC m=+1040.592968007" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.137358 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" podStartSLOduration=3.32219073 podStartE2EDuration="31.137341961s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.866059764 +0000 UTC m=+1011.354724386" lastFinishedPulling="2025-12-05 08:32:54.681210995 +0000 UTC m=+1039.169875617" observedRunningTime="2025-12-05 08:32:56.134694201 +0000 UTC m=+1040.623358823" watchObservedRunningTime="2025-12-05 08:32:56.137341961 +0000 UTC m=+1040.626006583" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.220124 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-v2p7b" podStartSLOduration=4.387746052 podStartE2EDuration="32.220103051s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.71437598 +0000 UTC m=+1011.203040602" lastFinishedPulling="2025-12-05 08:32:54.546732979 +0000 UTC m=+1039.035397601" observedRunningTime="2025-12-05 08:32:56.15905192 +0000 UTC m=+1040.647716542" watchObservedRunningTime="2025-12-05 08:32:56.220103051 +0000 UTC m=+1040.708767673" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.276612 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-jwn22" podStartSLOduration=4.023195191 podStartE2EDuration="32.276596762s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.158096863 +0000 UTC m=+1010.646761485" lastFinishedPulling="2025-12-05 08:32:54.411498434 +0000 UTC m=+1038.900163056" observedRunningTime="2025-12-05 08:32:56.272396322 +0000 UTC m=+1040.761060944" watchObservedRunningTime="2025-12-05 08:32:56.276596762 +0000 UTC m=+1040.765261384" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.314097 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" podStartSLOduration=3.865365521 podStartE2EDuration="31.314078804s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.740564925 +0000 UTC m=+1011.229229547" lastFinishedPulling="2025-12-05 08:32:54.189278208 +0000 UTC m=+1038.677942830" observedRunningTime="2025-12-05 08:32:56.308460497 +0000 UTC m=+1040.797125119" watchObservedRunningTime="2025-12-05 08:32:56.314078804 +0000 UTC m=+1040.802743426" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.350968 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lfncq" podStartSLOduration=3.69796905 podStartE2EDuration="31.350950531s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.732421039 +0000 UTC m=+1011.221085661" lastFinishedPulling="2025-12-05 08:32:54.38540252 +0000 UTC m=+1038.874067142" observedRunningTime="2025-12-05 08:32:56.348951369 +0000 UTC m=+1040.837616001" watchObservedRunningTime="2025-12-05 08:32:56.350950531 +0000 UTC m=+1040.839615153" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.925624 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:56 crc kubenswrapper[4876]: I1205 08:32:56.933781 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/679c172a-10ff-4a0e-8932-bcc1f78393d7-cert\") pod \"infra-operator-controller-manager-57548d458d-l9jr7\" (UID: \"679c172a-10ff-4a0e-8932-bcc1f78393d7\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.089996 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-qltmw" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.090038 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wptqj" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.090119 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-svc6v" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.090510 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-tn7mh" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.091420 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lxwt9" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.157319 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-49vd6" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.159176 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.242690 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.250877 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e3a38502-6fad-4383-81e8-591318a4d4d3-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6\" (UID: \"e3a38502-6fad-4383-81e8-591318a4d4d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.387548 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8llhg" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.396329 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.547171 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.547285 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.552456 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-webhook-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.552640 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48cca8bf-c0e1-4957-885a-4cf42c1c33fe-metrics-certs\") pod \"openstack-operator-controller-manager-8579f747db-9t4h4\" (UID: \"48cca8bf-c0e1-4957-885a-4cf42c1c33fe\") " pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.608335 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7"] Dec 05 08:32:57 crc kubenswrapper[4876]: W1205 08:32:57.610806 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod679c172a_10ff_4a0e_8932_bcc1f78393d7.slice/crio-7b817ec86e1c1bef7d80893478993c6286969a9bcc704ef47db901979586f666 WatchSource:0}: Error finding container 7b817ec86e1c1bef7d80893478993c6286969a9bcc704ef47db901979586f666: Status 404 returned error can't find the container with id 7b817ec86e1c1bef7d80893478993c6286969a9bcc704ef47db901979586f666 Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.641601 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6"] Dec 05 08:32:57 crc kubenswrapper[4876]: W1205 08:32:57.650274 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3a38502_6fad_4383_81e8_591318a4d4d3.slice/crio-6725648a75140d24eb3096d86fbaf41e7538710c05fcc1072df62e9a30bee118 WatchSource:0}: Error finding container 6725648a75140d24eb3096d86fbaf41e7538710c05fcc1072df62e9a30bee118: Status 404 returned error can't find the container with id 6725648a75140d24eb3096d86fbaf41e7538710c05fcc1072df62e9a30bee118 Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.719890 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-lqrqm" Dec 05 08:32:57 crc kubenswrapper[4876]: I1205 08:32:57.728292 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:32:58 crc kubenswrapper[4876]: I1205 08:32:58.095397 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" event={"ID":"e3a38502-6fad-4383-81e8-591318a4d4d3","Type":"ContainerStarted","Data":"6725648a75140d24eb3096d86fbaf41e7538710c05fcc1072df62e9a30bee118"} Dec 05 08:32:58 crc kubenswrapper[4876]: I1205 08:32:58.097229 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" event={"ID":"679c172a-10ff-4a0e-8932-bcc1f78393d7","Type":"ContainerStarted","Data":"7b817ec86e1c1bef7d80893478993c6286969a9bcc704ef47db901979586f666"} Dec 05 08:32:58 crc kubenswrapper[4876]: I1205 08:32:58.125643 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4"] Dec 05 08:32:58 crc kubenswrapper[4876]: W1205 08:32:58.134543 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48cca8bf_c0e1_4957_885a_4cf42c1c33fe.slice/crio-fac65235743bd44b5d13e2196e26a5cd2200d3c0919a2833873255723bdc38e8 WatchSource:0}: Error finding container fac65235743bd44b5d13e2196e26a5cd2200d3c0919a2833873255723bdc38e8: Status 404 returned error can't find the container with id fac65235743bd44b5d13e2196e26a5cd2200d3c0919a2833873255723bdc38e8 Dec 05 08:32:59 crc kubenswrapper[4876]: I1205 08:32:59.106411 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" event={"ID":"48cca8bf-c0e1-4957-885a-4cf42c1c33fe","Type":"ContainerStarted","Data":"fac65235743bd44b5d13e2196e26a5cd2200d3c0919a2833873255723bdc38e8"} Dec 05 08:33:05 crc kubenswrapper[4876]: I1205 08:33:05.833129 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-hxc5v" Dec 05 08:33:07 crc kubenswrapper[4876]: I1205 08:33:07.248979 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" event={"ID":"0b8ca916-a8c5-4866-8333-ebbc0f861161","Type":"ContainerStarted","Data":"59e491d9168ab918013af526ac1b466408eda2397e64765d2ed4dfd098ef1434"} Dec 05 08:33:08 crc kubenswrapper[4876]: I1205 08:33:08.213589 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:33:08 crc kubenswrapper[4876]: I1205 08:33:08.213935 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:33:08 crc kubenswrapper[4876]: I1205 08:33:08.257642 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" event={"ID":"e21b6aa8-f703-42e0-8251-06444e6423fe","Type":"ContainerStarted","Data":"e7ecfe515f8316b44800715ec8ba3c8ef56c7bd5575af825f937b5ec01a5ca57"} Dec 05 08:33:10 crc kubenswrapper[4876]: I1205 08:33:10.276228 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" event={"ID":"48cca8bf-c0e1-4957-885a-4cf42c1c33fe","Type":"ContainerStarted","Data":"f1bade63b235aaf52d3443ef07349e856d4c1313780b5eb0122690dbb5197aba"} Dec 05 08:33:10 crc kubenswrapper[4876]: I1205 08:33:10.281019 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" event={"ID":"13b56ee6-b8fa-40bf-add8-1778f4acf30d","Type":"ContainerStarted","Data":"87323ebc060b959369c063d90a84c628adeafbd4981ce73053112d7c9e03c483"} Dec 05 08:33:10 crc kubenswrapper[4876]: I1205 08:33:10.286404 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" event={"ID":"68006b91-3398-4ff1-8c6e-ede02835c0ef","Type":"ContainerStarted","Data":"ab4e95752a483db157df073a54d6365730842639bccdc1184ec007dd59096b82"} Dec 05 08:33:10 crc kubenswrapper[4876]: I1205 08:33:10.316386 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" podStartSLOduration=17.493732728 podStartE2EDuration="46.316368962s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.739670471 +0000 UTC m=+1011.228335093" lastFinishedPulling="2025-12-05 08:32:55.562306705 +0000 UTC m=+1040.050971327" observedRunningTime="2025-12-05 08:33:10.310352454 +0000 UTC m=+1054.799017076" watchObservedRunningTime="2025-12-05 08:33:10.316368962 +0000 UTC m=+1054.805033584" Dec 05 08:33:11 crc kubenswrapper[4876]: I1205 08:33:11.295077 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:33:11 crc kubenswrapper[4876]: I1205 08:33:11.295375 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" Dec 05 08:33:11 crc kubenswrapper[4876]: I1205 08:33:11.295493 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" Dec 05 08:33:11 crc kubenswrapper[4876]: I1205 08:33:11.308147 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" podStartSLOduration=18.452984897 podStartE2EDuration="47.308128072s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.713711642 +0000 UTC m=+1011.202376264" lastFinishedPulling="2025-12-05 08:32:55.568854807 +0000 UTC m=+1040.057519439" observedRunningTime="2025-12-05 08:33:11.307153837 +0000 UTC m=+1055.795818459" watchObservedRunningTime="2025-12-05 08:33:11.308128072 +0000 UTC m=+1055.796792694" Dec 05 08:33:11 crc kubenswrapper[4876]: I1205 08:33:11.337844 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" podStartSLOduration=46.337829771 podStartE2EDuration="46.337829771s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:33:11.334200796 +0000 UTC m=+1055.822865418" watchObservedRunningTime="2025-12-05 08:33:11.337829771 +0000 UTC m=+1055.826494393" Dec 05 08:33:11 crc kubenswrapper[4876]: I1205 08:33:11.371780 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" podStartSLOduration=17.513794733 podStartE2EDuration="46.371759331s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.714173135 +0000 UTC m=+1011.202837757" lastFinishedPulling="2025-12-05 08:32:55.572137723 +0000 UTC m=+1040.060802355" observedRunningTime="2025-12-05 08:33:11.363192056 +0000 UTC m=+1055.851856678" watchObservedRunningTime="2025-12-05 08:33:11.371759331 +0000 UTC m=+1055.860423953" Dec 05 08:33:11 crc kubenswrapper[4876]: I1205 08:33:11.393658 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" podStartSLOduration=17.5638388 podStartE2EDuration="46.393639244s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.732167362 +0000 UTC m=+1011.220831984" lastFinishedPulling="2025-12-05 08:32:55.561967796 +0000 UTC m=+1040.050632428" observedRunningTime="2025-12-05 08:33:11.389164307 +0000 UTC m=+1055.877828929" watchObservedRunningTime="2025-12-05 08:33:11.393639244 +0000 UTC m=+1055.882303866" Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.314702 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" event={"ID":"ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7","Type":"ContainerStarted","Data":"5c234136234ee987b2c954f29fcb963592af37d31a2fa192ace60b41d73c169f"} Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.315472 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.316887 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" event={"ID":"679c172a-10ff-4a0e-8932-bcc1f78393d7","Type":"ContainerStarted","Data":"75dde67477253e80d300885aff9dfe3af952c8bece48740683c98a1359344905"} Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.316960 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" event={"ID":"679c172a-10ff-4a0e-8932-bcc1f78393d7","Type":"ContainerStarted","Data":"f748491820abb984c2e49856bc4d6ba8724aeead0e825c79baad9eccc57d7155"} Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.317073 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.318803 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" event={"ID":"e3a38502-6fad-4383-81e8-591318a4d4d3","Type":"ContainerStarted","Data":"5671f739088eda9456a35981d96b59393eb81a3b1c83dbed9c99069e05a059d4"} Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.318851 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" event={"ID":"e3a38502-6fad-4383-81e8-591318a4d4d3","Type":"ContainerStarted","Data":"6c64f4080590bf1b10e6a24c838c88ff5360ecb2bc5361136bff266aaeefca5e"} Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.319047 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.331704 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" podStartSLOduration=4.016348633 podStartE2EDuration="50.331678241s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:26.713941848 +0000 UTC m=+1011.202606470" lastFinishedPulling="2025-12-05 08:33:13.029271456 +0000 UTC m=+1057.517936078" observedRunningTime="2025-12-05 08:33:14.330805798 +0000 UTC m=+1058.819470430" watchObservedRunningTime="2025-12-05 08:33:14.331678241 +0000 UTC m=+1058.820342903" Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.363941 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" podStartSLOduration=33.982213254 podStartE2EDuration="49.363919116s" podCreationTimestamp="2025-12-05 08:32:25 +0000 UTC" firstStartedPulling="2025-12-05 08:32:57.652683168 +0000 UTC m=+1042.141347790" lastFinishedPulling="2025-12-05 08:33:13.03438903 +0000 UTC m=+1057.523053652" observedRunningTime="2025-12-05 08:33:14.358034942 +0000 UTC m=+1058.846699564" watchObservedRunningTime="2025-12-05 08:33:14.363919116 +0000 UTC m=+1058.852583738" Dec 05 08:33:14 crc kubenswrapper[4876]: I1205 08:33:14.383204 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" podStartSLOduration=34.95383738 podStartE2EDuration="50.383183911s" podCreationTimestamp="2025-12-05 08:32:24 +0000 UTC" firstStartedPulling="2025-12-05 08:32:57.615106413 +0000 UTC m=+1042.103771045" lastFinishedPulling="2025-12-05 08:33:13.044452954 +0000 UTC m=+1057.533117576" observedRunningTime="2025-12-05 08:33:14.380253095 +0000 UTC m=+1058.868917717" watchObservedRunningTime="2025-12-05 08:33:14.383183911 +0000 UTC m=+1058.871848533" Dec 05 08:33:15 crc kubenswrapper[4876]: I1205 08:33:15.331692 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" Dec 05 08:33:15 crc kubenswrapper[4876]: I1205 08:33:15.333778 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vjh2v" Dec 05 08:33:15 crc kubenswrapper[4876]: I1205 08:33:15.370413 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" Dec 05 08:33:15 crc kubenswrapper[4876]: I1205 08:33:15.391300 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ckhg" Dec 05 08:33:15 crc kubenswrapper[4876]: I1205 08:33:15.419741 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-vp2x4" Dec 05 08:33:15 crc kubenswrapper[4876]: I1205 08:33:15.487480 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2spcs" Dec 05 08:33:17 crc kubenswrapper[4876]: I1205 08:33:17.735863 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8579f747db-9t4h4" Dec 05 08:33:25 crc kubenswrapper[4876]: I1205 08:33:25.362629 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-qgxmr" Dec 05 08:33:27 crc kubenswrapper[4876]: I1205 08:33:27.170111 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-l9jr7" Dec 05 08:33:27 crc kubenswrapper[4876]: I1205 08:33:27.402703 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6" Dec 05 08:33:38 crc kubenswrapper[4876]: I1205 08:33:38.213596 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:33:38 crc kubenswrapper[4876]: I1205 08:33:38.215072 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.608401 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9qg"] Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.611818 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.615076 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8hbkd" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.615614 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.615678 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.615679 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.636345 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9qg"] Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.767555 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfmfn\" (UniqueName: \"kubernetes.io/projected/7e95067d-f184-4315-8be4-0cb9a115b379-kube-api-access-hfmfn\") pod \"dnsmasq-dns-675f4bcbfc-xh9qg\" (UID: \"7e95067d-f184-4315-8be4-0cb9a115b379\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.767743 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e95067d-f184-4315-8be4-0cb9a115b379-config\") pod \"dnsmasq-dns-675f4bcbfc-xh9qg\" (UID: \"7e95067d-f184-4315-8be4-0cb9a115b379\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.852224 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-86m8k"] Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.853314 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.856865 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.876864 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e95067d-f184-4315-8be4-0cb9a115b379-config\") pod \"dnsmasq-dns-675f4bcbfc-xh9qg\" (UID: \"7e95067d-f184-4315-8be4-0cb9a115b379\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.876935 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfmfn\" (UniqueName: \"kubernetes.io/projected/7e95067d-f184-4315-8be4-0cb9a115b379-kube-api-access-hfmfn\") pod \"dnsmasq-dns-675f4bcbfc-xh9qg\" (UID: \"7e95067d-f184-4315-8be4-0cb9a115b379\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.878025 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e95067d-f184-4315-8be4-0cb9a115b379-config\") pod \"dnsmasq-dns-675f4bcbfc-xh9qg\" (UID: \"7e95067d-f184-4315-8be4-0cb9a115b379\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.881143 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-86m8k"] Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.916753 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfmfn\" (UniqueName: \"kubernetes.io/projected/7e95067d-f184-4315-8be4-0cb9a115b379-kube-api-access-hfmfn\") pod \"dnsmasq-dns-675f4bcbfc-xh9qg\" (UID: \"7e95067d-f184-4315-8be4-0cb9a115b379\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.938463 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.981754 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-config\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.981809 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:42 crc kubenswrapper[4876]: I1205 08:33:42.981852 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv57z\" (UniqueName: \"kubernetes.io/projected/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-kube-api-access-rv57z\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.084500 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.085799 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.085856 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv57z\" (UniqueName: \"kubernetes.io/projected/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-kube-api-access-rv57z\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.085988 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-config\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.086612 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-config\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.106489 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv57z\" (UniqueName: \"kubernetes.io/projected/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-kube-api-access-rv57z\") pod \"dnsmasq-dns-78dd6ddcc-86m8k\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.168302 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.487929 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9qg"] Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.495012 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:33:43 crc kubenswrapper[4876]: I1205 08:33:43.599441 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-86m8k"] Dec 05 08:33:43 crc kubenswrapper[4876]: W1205 08:33:43.603274 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1f67237_8c66_4107_a8fb_e3e8eb8b89ca.slice/crio-55bbf34d4d5c8f9b3f76956d20d86a25c4d81340ae2771679fee0dd0cdd61f76 WatchSource:0}: Error finding container 55bbf34d4d5c8f9b3f76956d20d86a25c4d81340ae2771679fee0dd0cdd61f76: Status 404 returned error can't find the container with id 55bbf34d4d5c8f9b3f76956d20d86a25c4d81340ae2771679fee0dd0cdd61f76 Dec 05 08:33:44 crc kubenswrapper[4876]: I1205 08:33:44.007409 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" event={"ID":"7e95067d-f184-4315-8be4-0cb9a115b379","Type":"ContainerStarted","Data":"32e3ed970ab432ea4ccffd5c7671a5fb8e053514b26922f5284dba1a8e23d637"} Dec 05 08:33:44 crc kubenswrapper[4876]: I1205 08:33:44.009201 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" event={"ID":"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca","Type":"ContainerStarted","Data":"55bbf34d4d5c8f9b3f76956d20d86a25c4d81340ae2771679fee0dd0cdd61f76"} Dec 05 08:33:45 crc kubenswrapper[4876]: I1205 08:33:45.791731 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9qg"] Dec 05 08:33:45 crc kubenswrapper[4876]: I1205 08:33:45.839172 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x2c7b"] Dec 05 08:33:45 crc kubenswrapper[4876]: I1205 08:33:45.840511 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x2c7b"] Dec 05 08:33:45 crc kubenswrapper[4876]: I1205 08:33:45.840690 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:45 crc kubenswrapper[4876]: I1205 08:33:45.931699 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-672ft\" (UniqueName: \"kubernetes.io/projected/03da7075-fc4f-42e1-bc60-0acdf6fef427-kube-api-access-672ft\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:45 crc kubenswrapper[4876]: I1205 08:33:45.931743 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-dns-svc\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:45 crc kubenswrapper[4876]: I1205 08:33:45.931766 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-config\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.032579 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-672ft\" (UniqueName: \"kubernetes.io/projected/03da7075-fc4f-42e1-bc60-0acdf6fef427-kube-api-access-672ft\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.032892 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-dns-svc\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.032932 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-config\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.033864 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-config\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.034653 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-dns-svc\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.066751 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-672ft\" (UniqueName: \"kubernetes.io/projected/03da7075-fc4f-42e1-bc60-0acdf6fef427-kube-api-access-672ft\") pod \"dnsmasq-dns-666b6646f7-x2c7b\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.168525 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.219549 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-86m8k"] Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.258274 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dtnkx"] Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.271597 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.271968 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dtnkx"] Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.349734 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv9p2\" (UniqueName: \"kubernetes.io/projected/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-kube-api-access-wv9p2\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.349799 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.349855 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-config\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.450598 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-config\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.450875 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv9p2\" (UniqueName: \"kubernetes.io/projected/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-kube-api-access-wv9p2\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.450936 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.452158 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.452215 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-config\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.483317 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv9p2\" (UniqueName: \"kubernetes.io/projected/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-kube-api-access-wv9p2\") pod \"dnsmasq-dns-57d769cc4f-dtnkx\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.546487 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x2c7b"] Dec 05 08:33:46 crc kubenswrapper[4876]: I1205 08:33:46.592586 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.058387 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" event={"ID":"03da7075-fc4f-42e1-bc60-0acdf6fef427","Type":"ContainerStarted","Data":"c3206172c9e527a4f9f9136cee66f8b152ce3bdc3d2d7a3429da15939e701b48"} Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.068956 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.076105 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.080036 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.080101 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.084283 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.084425 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.084578 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.084997 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.085193 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h2wdn" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.117744 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.255968 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256577 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256650 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256675 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-config-data\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256689 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256734 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256770 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256802 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.256830 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.257098 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkq9s\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-kube-api-access-qkq9s\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.286059 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dtnkx"] Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358188 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358239 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358269 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358294 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358315 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358351 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkq9s\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-kube-api-access-qkq9s\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358370 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358411 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358426 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358445 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-config-data\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.358460 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.359443 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.359961 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.360318 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.360750 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.363374 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.363890 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.370914 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.378040 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-config-data\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.384241 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.385646 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.391230 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.391446 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.391554 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.391631 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.391762 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.391560 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.395480 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkq9s\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-kube-api-access-qkq9s\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.405317 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mmjr5" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.409435 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.410743 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.417467 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.429473 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.436368 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565026 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rljh\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-kube-api-access-4rljh\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565093 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565113 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565136 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565165 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565190 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565215 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565236 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565265 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565288 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/45b18655-19ec-43f4-a779-b8f72bc77aa2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.565329 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/45b18655-19ec-43f4-a779-b8f72bc77aa2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672501 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/45b18655-19ec-43f4-a779-b8f72bc77aa2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672646 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rljh\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-kube-api-access-4rljh\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672670 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672718 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672738 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672760 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672775 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672795 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672810 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672833 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.672852 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/45b18655-19ec-43f4-a779-b8f72bc77aa2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.676297 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.924770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.925021 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.926034 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.930723 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.931369 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.932649 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/45b18655-19ec-43f4-a779-b8f72bc77aa2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.933975 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.934408 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/45b18655-19ec-43f4-a779-b8f72bc77aa2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.939574 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.940119 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rljh\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-kube-api-access-4rljh\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:47 crc kubenswrapper[4876]: I1205 08:33:47.963620 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.088316 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.089171 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" event={"ID":"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6","Type":"ContainerStarted","Data":"64550113a8561e8fedef1ff37e0f739c6631ee31df88fe0db26bbc62f79baafa"} Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.210858 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.892132 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.929817 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.929972 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.933101 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.933567 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-qq478" Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.933980 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.934111 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.965541 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 05 08:33:48 crc kubenswrapper[4876]: I1205 08:33:48.999770 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:48.999992 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-kolla-config\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.000082 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.000113 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.000141 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22qxl\" (UniqueName: \"kubernetes.io/projected/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-kube-api-access-22qxl\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.000205 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-config-data-default\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.000227 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.000269 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.047044 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.104712 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.107511 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-kolla-config\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.107622 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.107645 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.107684 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22qxl\" (UniqueName: \"kubernetes.io/projected/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-kube-api-access-22qxl\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.107763 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-config-data-default\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.107832 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.107859 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.108551 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-kolla-config\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.109291 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.111197 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f","Type":"ContainerStarted","Data":"ac3247a496e8447132b39c1b5aa4f3e18d81a65ca345b38845a7e18d7235214d"} Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.112878 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-config-data-default\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.113247 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.121230 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.136105 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.136782 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.149277 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22qxl\" (UniqueName: \"kubernetes.io/projected/d47eff06-2977-4bcb-bc3b-6ad08399bbd7-kube-api-access-22qxl\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.188122 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"d47eff06-2977-4bcb-bc3b-6ad08399bbd7\") " pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.282199 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 08:33:49 crc kubenswrapper[4876]: I1205 08:33:49.883222 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.134351 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d47eff06-2977-4bcb-bc3b-6ad08399bbd7","Type":"ContainerStarted","Data":"cd54c5758ee43728aa131ac720e1ecf35d544125b51f7eb1b7a88ca0bd5c53f9"} Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.136519 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45b18655-19ec-43f4-a779-b8f72bc77aa2","Type":"ContainerStarted","Data":"56c55451e074fb1f9579db7475e31ac8e5c00ad401cb11503154d025e630feaf"} Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.464354 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.467429 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.474109 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-bl56b" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.474939 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.475128 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.475247 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.482958 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.502766 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.503985 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.510386 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.510624 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.511459 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-59lkt" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.540787 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.545742 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/761374dd-700e-42fa-81d4-d560f97643be-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.545807 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/846c521d-f570-4731-a923-8a1e9c626659-config-data\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.545834 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846c521d-f570-4731-a923-8a1e9c626659-combined-ca-bundle\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546141 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546169 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546192 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761374dd-700e-42fa-81d4-d560f97643be-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546218 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/761374dd-700e-42fa-81d4-d560f97643be-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546240 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546255 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/846c521d-f570-4731-a923-8a1e9c626659-kolla-config\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546287 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt5pm\" (UniqueName: \"kubernetes.io/projected/846c521d-f570-4731-a923-8a1e9c626659-kube-api-access-kt5pm\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546324 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdn6x\" (UniqueName: \"kubernetes.io/projected/761374dd-700e-42fa-81d4-d560f97643be-kube-api-access-rdn6x\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546341 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/846c521d-f570-4731-a923-8a1e9c626659-memcached-tls-certs\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.546372 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.647915 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/761374dd-700e-42fa-81d4-d560f97643be-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648144 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648166 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/846c521d-f570-4731-a923-8a1e9c626659-kolla-config\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648203 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt5pm\" (UniqueName: \"kubernetes.io/projected/846c521d-f570-4731-a923-8a1e9c626659-kube-api-access-kt5pm\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648260 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdn6x\" (UniqueName: \"kubernetes.io/projected/761374dd-700e-42fa-81d4-d560f97643be-kube-api-access-rdn6x\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648279 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/846c521d-f570-4731-a923-8a1e9c626659-memcached-tls-certs\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648307 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648329 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/761374dd-700e-42fa-81d4-d560f97643be-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648357 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/846c521d-f570-4731-a923-8a1e9c626659-config-data\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648376 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648392 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846c521d-f570-4731-a923-8a1e9c626659-combined-ca-bundle\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648423 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.648442 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761374dd-700e-42fa-81d4-d560f97643be-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.652801 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/846c521d-f570-4731-a923-8a1e9c626659-config-data\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.653455 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.655009 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/761374dd-700e-42fa-81d4-d560f97643be-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.656595 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.656825 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/846c521d-f570-4731-a923-8a1e9c626659-kolla-config\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.657360 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.657724 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/761374dd-700e-42fa-81d4-d560f97643be-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.665302 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761374dd-700e-42fa-81d4-d560f97643be-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.671630 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/846c521d-f570-4731-a923-8a1e9c626659-combined-ca-bundle\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.674794 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/846c521d-f570-4731-a923-8a1e9c626659-memcached-tls-certs\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.691589 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.693259 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdn6x\" (UniqueName: \"kubernetes.io/projected/761374dd-700e-42fa-81d4-d560f97643be-kube-api-access-rdn6x\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.693912 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt5pm\" (UniqueName: \"kubernetes.io/projected/846c521d-f570-4731-a923-8a1e9c626659-kube-api-access-kt5pm\") pod \"memcached-0\" (UID: \"846c521d-f570-4731-a923-8a1e9c626659\") " pod="openstack/memcached-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.694931 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/761374dd-700e-42fa-81d4-d560f97643be-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"761374dd-700e-42fa-81d4-d560f97643be\") " pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.827798 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 08:33:50 crc kubenswrapper[4876]: I1205 08:33:50.855143 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 08:33:52 crc kubenswrapper[4876]: I1205 08:33:52.451138 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:33:52 crc kubenswrapper[4876]: I1205 08:33:52.455709 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 08:33:52 crc kubenswrapper[4876]: I1205 08:33:52.459872 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-srhzt" Dec 05 08:33:52 crc kubenswrapper[4876]: I1205 08:33:52.485488 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:33:52 crc kubenswrapper[4876]: I1205 08:33:52.545134 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klg97\" (UniqueName: \"kubernetes.io/projected/b29eddc0-b7dd-4390-ab85-0e8eb61422e2-kube-api-access-klg97\") pod \"kube-state-metrics-0\" (UID: \"b29eddc0-b7dd-4390-ab85-0e8eb61422e2\") " pod="openstack/kube-state-metrics-0" Dec 05 08:33:52 crc kubenswrapper[4876]: I1205 08:33:52.646584 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klg97\" (UniqueName: \"kubernetes.io/projected/b29eddc0-b7dd-4390-ab85-0e8eb61422e2-kube-api-access-klg97\") pod \"kube-state-metrics-0\" (UID: \"b29eddc0-b7dd-4390-ab85-0e8eb61422e2\") " pod="openstack/kube-state-metrics-0" Dec 05 08:33:52 crc kubenswrapper[4876]: I1205 08:33:52.686569 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klg97\" (UniqueName: \"kubernetes.io/projected/b29eddc0-b7dd-4390-ab85-0e8eb61422e2-kube-api-access-klg97\") pod \"kube-state-metrics-0\" (UID: \"b29eddc0-b7dd-4390-ab85-0e8eb61422e2\") " pod="openstack/kube-state-metrics-0" Dec 05 08:33:52 crc kubenswrapper[4876]: I1205 08:33:52.788629 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.440820 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7p5wk"] Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.442301 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.445269 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-jjkkl" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.445429 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.446569 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.465410 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-fhwc7"] Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.467152 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.472933 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7p5wk"] Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.484987 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fhwc7"] Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.600952 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-run\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.600994 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07839a94-e94b-4ff2-9eaf-9fb315c83046-scripts\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601021 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-scripts\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601055 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-etc-ovs\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601073 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-log-ovn\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601094 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8m42\" (UniqueName: \"kubernetes.io/projected/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-kube-api-access-m8m42\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601240 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-run-ovn\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601362 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-combined-ca-bundle\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601436 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzkpk\" (UniqueName: \"kubernetes.io/projected/07839a94-e94b-4ff2-9eaf-9fb315c83046-kube-api-access-vzkpk\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601490 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-ovn-controller-tls-certs\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601552 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-run\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601644 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-log\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.601706 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-lib\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703611 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-log-ovn\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703667 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8m42\" (UniqueName: \"kubernetes.io/projected/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-kube-api-access-m8m42\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703716 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-run-ovn\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703762 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-combined-ca-bundle\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703803 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzkpk\" (UniqueName: \"kubernetes.io/projected/07839a94-e94b-4ff2-9eaf-9fb315c83046-kube-api-access-vzkpk\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703835 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-ovn-controller-tls-certs\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703868 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-run\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703918 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-log\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703949 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-lib\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.703981 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-run\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704005 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07839a94-e94b-4ff2-9eaf-9fb315c83046-scripts\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704028 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-scripts\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-etc-ovs\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704355 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-run\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704358 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-run\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704408 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-etc-ovs\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704512 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-log\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704702 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/07839a94-e94b-4ff2-9eaf-9fb315c83046-var-lib\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.704851 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-run-ovn\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.705653 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-var-log-ovn\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.710503 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-scripts\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.713538 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-ovn-controller-tls-certs\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.713770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-combined-ca-bundle\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.718178 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07839a94-e94b-4ff2-9eaf-9fb315c83046-scripts\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.720063 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8m42\" (UniqueName: \"kubernetes.io/projected/442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2-kube-api-access-m8m42\") pod \"ovn-controller-7p5wk\" (UID: \"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2\") " pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.736885 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzkpk\" (UniqueName: \"kubernetes.io/projected/07839a94-e94b-4ff2-9eaf-9fb315c83046-kube-api-access-vzkpk\") pod \"ovn-controller-ovs-fhwc7\" (UID: \"07839a94-e94b-4ff2-9eaf-9fb315c83046\") " pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.764733 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7p5wk" Dec 05 08:33:55 crc kubenswrapper[4876]: I1205 08:33:55.792471 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.035230 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.036526 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.038699 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.038994 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.039407 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-fn6jk" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.039723 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.039871 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.070125 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.115117 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.115168 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f18961c-c654-4f94-98a3-697644fdfda4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.115299 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.115328 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbj4q\" (UniqueName: \"kubernetes.io/projected/3f18961c-c654-4f94-98a3-697644fdfda4-kube-api-access-xbj4q\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.115348 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f18961c-c654-4f94-98a3-697644fdfda4-config\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.115366 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3f18961c-c654-4f94-98a3-697644fdfda4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.115387 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.115417 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.220383 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.220486 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.220527 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f18961c-c654-4f94-98a3-697644fdfda4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.220579 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.220616 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbj4q\" (UniqueName: \"kubernetes.io/projected/3f18961c-c654-4f94-98a3-697644fdfda4-kube-api-access-xbj4q\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.220638 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f18961c-c654-4f94-98a3-697644fdfda4-config\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.220659 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3f18961c-c654-4f94-98a3-697644fdfda4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.220686 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.222673 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f18961c-c654-4f94-98a3-697644fdfda4-config\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.225410 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.234524 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.238451 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.245744 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3f18961c-c654-4f94-98a3-697644fdfda4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.246301 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f18961c-c654-4f94-98a3-697644fdfda4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.248563 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f18961c-c654-4f94-98a3-697644fdfda4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.256503 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbj4q\" (UniqueName: \"kubernetes.io/projected/3f18961c-c654-4f94-98a3-697644fdfda4-kube-api-access-xbj4q\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.279005 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3f18961c-c654-4f94-98a3-697644fdfda4\") " pod="openstack/ovsdbserver-nb-0" Dec 05 08:33:56 crc kubenswrapper[4876]: I1205 08:33:56.351819 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.111965 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.113371 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.114991 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.115598 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.115647 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.115775 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-szx7c" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.124479 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.200351 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c3699c2-facd-4d82-92ec-4816b871833b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.200431 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.200462 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.200487 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm9h4\" (UniqueName: \"kubernetes.io/projected/9c3699c2-facd-4d82-92ec-4816b871833b-kube-api-access-tm9h4\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.200704 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.200747 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c3699c2-facd-4d82-92ec-4816b871833b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.200973 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.201153 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c3699c2-facd-4d82-92ec-4816b871833b-config\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.302748 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c3699c2-facd-4d82-92ec-4816b871833b-config\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.302816 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c3699c2-facd-4d82-92ec-4816b871833b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.302875 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.302985 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.303025 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm9h4\" (UniqueName: \"kubernetes.io/projected/9c3699c2-facd-4d82-92ec-4816b871833b-kube-api-access-tm9h4\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.303058 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.303088 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c3699c2-facd-4d82-92ec-4816b871833b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.303165 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.303526 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.304018 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c3699c2-facd-4d82-92ec-4816b871833b-config\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.304367 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c3699c2-facd-4d82-92ec-4816b871833b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.310958 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.311239 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c3699c2-facd-4d82-92ec-4816b871833b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.313744 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.389021 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3699c2-facd-4d82-92ec-4816b871833b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.389838 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm9h4\" (UniqueName: \"kubernetes.io/projected/9c3699c2-facd-4d82-92ec-4816b871833b-kube-api-access-tm9h4\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.433419 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9c3699c2-facd-4d82-92ec-4816b871833b\") " pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:00 crc kubenswrapper[4876]: I1205 08:34:00.730749 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:08 crc kubenswrapper[4876]: I1205 08:34:08.213569 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:34:08 crc kubenswrapper[4876]: I1205 08:34:08.214130 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:34:08 crc kubenswrapper[4876]: I1205 08:34:08.214208 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:34:08 crc kubenswrapper[4876]: I1205 08:34:08.215254 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6167cbd18c38bbbdb4c177b56aa429711e7d69fa0ae0c347c0b78af72fe60dcb"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:34:08 crc kubenswrapper[4876]: I1205 08:34:08.215363 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://6167cbd18c38bbbdb4c177b56aa429711e7d69fa0ae0c347c0b78af72fe60dcb" gracePeriod=600 Dec 05 08:34:11 crc kubenswrapper[4876]: I1205 08:34:11.388187 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="6167cbd18c38bbbdb4c177b56aa429711e7d69fa0ae0c347c0b78af72fe60dcb" exitCode=0 Dec 05 08:34:11 crc kubenswrapper[4876]: I1205 08:34:11.388271 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"6167cbd18c38bbbdb4c177b56aa429711e7d69fa0ae0c347c0b78af72fe60dcb"} Dec 05 08:34:11 crc kubenswrapper[4876]: I1205 08:34:11.388820 4876 scope.go:117] "RemoveContainer" containerID="b93e463bd75bc474356eb877c940a3ca93130dc5e558c7d99bfe4997ecfbcedb" Dec 05 08:34:12 crc kubenswrapper[4876]: E1205 08:34:12.185498 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 05 08:34:12 crc kubenswrapper[4876]: E1205 08:34:12.185722 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4rljh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(45b18655-19ec-43f4-a779-b8f72bc77aa2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:34:12 crc kubenswrapper[4876]: E1205 08:34:12.186892 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" Dec 05 08:34:12 crc kubenswrapper[4876]: E1205 08:34:12.398029 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" Dec 05 08:34:21 crc kubenswrapper[4876]: E1205 08:34:21.677308 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 05 08:34:21 crc kubenswrapper[4876]: E1205 08:34:21.678165 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-22qxl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(d47eff06-2977-4bcb-bc3b-6ad08399bbd7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:34:21 crc kubenswrapper[4876]: E1205 08:34:21.680084 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="d47eff06-2977-4bcb-bc3b-6ad08399bbd7" Dec 05 08:34:21 crc kubenswrapper[4876]: E1205 08:34:21.695500 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 05 08:34:21 crc kubenswrapper[4876]: E1205 08:34:21.695659 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qkq9s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(7edf4ad5-0a62-4a2d-af4a-aecf59072c8f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:34:21 crc kubenswrapper[4876]: E1205 08:34:21.697693 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.409197 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.409648 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-672ft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-x2c7b_openstack(03da7075-fc4f-42e1-bc60-0acdf6fef427): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.410949 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" podUID="03da7075-fc4f-42e1-bc60-0acdf6fef427" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.430788 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.431012 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hfmfn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-xh9qg_openstack(7e95067d-f184-4315-8be4-0cb9a115b379): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.432243 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" podUID="7e95067d-f184-4315-8be4-0cb9a115b379" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.439648 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.439861 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wv9p2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-dtnkx_openstack(a6e5de03-4ce5-4c61-8589-7cd403a5e7a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.441354 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" podUID="a6e5de03-4ce5-4c61-8589-7cd403a5e7a6" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.474020 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.474227 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rv57z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-86m8k_openstack(b1f67237-8c66-4107-a8fb-e3e8eb8b89ca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.475389 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" podUID="b1f67237-8c66-4107-a8fb-e3e8eb8b89ca" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.537204 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="d47eff06-2977-4bcb-bc3b-6ad08399bbd7" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.537230 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.537310 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" podUID="03da7075-fc4f-42e1-bc60-0acdf6fef427" Dec 05 08:34:22 crc kubenswrapper[4876]: E1205 08:34:22.537344 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" podUID="a6e5de03-4ce5-4c61-8589-7cd403a5e7a6" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.073189 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.078862 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.142419 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.146661 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.163482 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.194434 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.230140 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfmfn\" (UniqueName: \"kubernetes.io/projected/7e95067d-f184-4315-8be4-0cb9a115b379-kube-api-access-hfmfn\") pod \"7e95067d-f184-4315-8be4-0cb9a115b379\" (UID: \"7e95067d-f184-4315-8be4-0cb9a115b379\") " Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.230170 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-config\") pod \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.230187 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-dns-svc\") pod \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.230307 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv57z\" (UniqueName: \"kubernetes.io/projected/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-kube-api-access-rv57z\") pod \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\" (UID: \"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca\") " Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.230409 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e95067d-f184-4315-8be4-0cb9a115b379-config\") pod \"7e95067d-f184-4315-8be4-0cb9a115b379\" (UID: \"7e95067d-f184-4315-8be4-0cb9a115b379\") " Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.230734 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b1f67237-8c66-4107-a8fb-e3e8eb8b89ca" (UID: "b1f67237-8c66-4107-a8fb-e3e8eb8b89ca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.230996 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e95067d-f184-4315-8be4-0cb9a115b379-config" (OuterVolumeSpecName: "config") pod "7e95067d-f184-4315-8be4-0cb9a115b379" (UID: "7e95067d-f184-4315-8be4-0cb9a115b379"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.230641 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-config" (OuterVolumeSpecName: "config") pod "b1f67237-8c66-4107-a8fb-e3e8eb8b89ca" (UID: "b1f67237-8c66-4107-a8fb-e3e8eb8b89ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.235294 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e95067d-f184-4315-8be4-0cb9a115b379-kube-api-access-hfmfn" (OuterVolumeSpecName: "kube-api-access-hfmfn") pod "7e95067d-f184-4315-8be4-0cb9a115b379" (UID: "7e95067d-f184-4315-8be4-0cb9a115b379"). InnerVolumeSpecName "kube-api-access-hfmfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.235343 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-kube-api-access-rv57z" (OuterVolumeSpecName: "kube-api-access-rv57z") pod "b1f67237-8c66-4107-a8fb-e3e8eb8b89ca" (UID: "b1f67237-8c66-4107-a8fb-e3e8eb8b89ca"). InnerVolumeSpecName "kube-api-access-rv57z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:23 crc kubenswrapper[4876]: W1205 08:34:23.239999 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb29eddc0_b7dd_4390_ab85_0e8eb61422e2.slice/crio-75e0ed124f7c0f5d1179aa1d7f39781a3df329d3eb9d1b21201010653d86d0eb WatchSource:0}: Error finding container 75e0ed124f7c0f5d1179aa1d7f39781a3df329d3eb9d1b21201010653d86d0eb: Status 404 returned error can't find the container with id 75e0ed124f7c0f5d1179aa1d7f39781a3df329d3eb9d1b21201010653d86d0eb Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.268013 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 08:34:23 crc kubenswrapper[4876]: W1205 08:34:23.270688 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod442faaa2_6ae4_4b4a_aad9_fa3f8f157ba2.slice/crio-2ba6ef636800ff730d948453f8e02a56c63e029e64f96ccb506778205ff572b2 WatchSource:0}: Error finding container 2ba6ef636800ff730d948453f8e02a56c63e029e64f96ccb506778205ff572b2: Status 404 returned error can't find the container with id 2ba6ef636800ff730d948453f8e02a56c63e029e64f96ccb506778205ff572b2 Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.276641 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7p5wk"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.334941 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv57z\" (UniqueName: \"kubernetes.io/projected/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-kube-api-access-rv57z\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.335244 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e95067d-f184-4315-8be4-0cb9a115b379-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.335255 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfmfn\" (UniqueName: \"kubernetes.io/projected/7e95067d-f184-4315-8be4-0cb9a115b379-kube-api-access-hfmfn\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.335263 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.335273 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.386002 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-fhwc7"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.532828 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7p5wk" event={"ID":"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2","Type":"ContainerStarted","Data":"2ba6ef636800ff730d948453f8e02a56c63e029e64f96ccb506778205ff572b2"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.534621 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" event={"ID":"7e95067d-f184-4315-8be4-0cb9a115b379","Type":"ContainerDied","Data":"32e3ed970ab432ea4ccffd5c7671a5fb8e053514b26922f5284dba1a8e23d637"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.534679 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xh9qg" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.536266 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"761374dd-700e-42fa-81d4-d560f97643be","Type":"ContainerStarted","Data":"417056bbaa4a0ee02803c739fea311deec3edc08d9d866300f83af2a0e0ad785"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.538136 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" event={"ID":"b1f67237-8c66-4107-a8fb-e3e8eb8b89ca","Type":"ContainerDied","Data":"55bbf34d4d5c8f9b3f76956d20d86a25c4d81340ae2771679fee0dd0cdd61f76"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.538170 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-86m8k" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.539707 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhwc7" event={"ID":"07839a94-e94b-4ff2-9eaf-9fb315c83046","Type":"ContainerStarted","Data":"8fd42af241aef4068955231e4b1026a6383acc9e4a419e4381204014c8cba640"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.541214 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"846c521d-f570-4731-a923-8a1e9c626659","Type":"ContainerStarted","Data":"9e0685e586bdc4b66104193a4c328ad59fc175b1874c4e677686b78f3b4f081b"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.542399 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3f18961c-c654-4f94-98a3-697644fdfda4","Type":"ContainerStarted","Data":"74dc699459a75b3db54779a3647d4e90d134cb81d31b458dc96e8e869bfdba33"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.544403 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"9aa43e6caeca2c89f47c0cead848074edb893e13fe607a55570e6f144dca2a07"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.547563 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b29eddc0-b7dd-4390-ab85-0e8eb61422e2","Type":"ContainerStarted","Data":"75e0ed124f7c0f5d1179aa1d7f39781a3df329d3eb9d1b21201010653d86d0eb"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.548778 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9c3699c2-facd-4d82-92ec-4816b871833b","Type":"ContainerStarted","Data":"686e2d1bd6986b9ea3d0ecb385a25cee1294d9a1306b4d62f0b5e24e3f6672e1"} Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.598485 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9qg"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.615590 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xh9qg"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.637538 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-86m8k"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.643856 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-86m8k"] Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.837554 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e95067d-f184-4315-8be4-0cb9a115b379" path="/var/lib/kubelet/pods/7e95067d-f184-4315-8be4-0cb9a115b379/volumes" Dec 05 08:34:23 crc kubenswrapper[4876]: I1205 08:34:23.838136 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1f67237-8c66-4107-a8fb-e3e8eb8b89ca" path="/var/lib/kubelet/pods/b1f67237-8c66-4107-a8fb-e3e8eb8b89ca/volumes" Dec 05 08:34:24 crc kubenswrapper[4876]: I1205 08:34:24.581791 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"761374dd-700e-42fa-81d4-d560f97643be","Type":"ContainerStarted","Data":"5243bd858b9c742742451c5dae355cf118a01280bf0594cfc2191f7857c170be"} Dec 05 08:34:27 crc kubenswrapper[4876]: I1205 08:34:27.604371 4876 generic.go:334] "Generic (PLEG): container finished" podID="761374dd-700e-42fa-81d4-d560f97643be" containerID="5243bd858b9c742742451c5dae355cf118a01280bf0594cfc2191f7857c170be" exitCode=0 Dec 05 08:34:27 crc kubenswrapper[4876]: I1205 08:34:27.604490 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"761374dd-700e-42fa-81d4-d560f97643be","Type":"ContainerDied","Data":"5243bd858b9c742742451c5dae355cf118a01280bf0594cfc2191f7857c170be"} Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.621429 4876 generic.go:334] "Generic (PLEG): container finished" podID="07839a94-e94b-4ff2-9eaf-9fb315c83046" containerID="36ee3af422dc3de75d49eb914448bdccbe0709cf0b8ddca72b66d866bb2a9e57" exitCode=0 Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.621508 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhwc7" event={"ID":"07839a94-e94b-4ff2-9eaf-9fb315c83046","Type":"ContainerDied","Data":"36ee3af422dc3de75d49eb914448bdccbe0709cf0b8ddca72b66d866bb2a9e57"} Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.624261 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"846c521d-f570-4731-a923-8a1e9c626659","Type":"ContainerStarted","Data":"6fbf9c9cc15ce2d832d10d92662f7a388289713c8edc8a2927bb853c394707c4"} Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.624480 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.628033 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b29eddc0-b7dd-4390-ab85-0e8eb61422e2","Type":"ContainerStarted","Data":"4a541af3fa0c88410e322db126e7852014165b17d79a5029bf58c38739650f88"} Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.628173 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.634196 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9c3699c2-facd-4d82-92ec-4816b871833b","Type":"ContainerStarted","Data":"60c52cba06ddab734ce31e80178b59a397d860b6a85321cb475d0e70253224a4"} Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.636390 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7p5wk" event={"ID":"442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2","Type":"ContainerStarted","Data":"e06d1d27a498f3fb7628b059f749c9b2750ff411c677c010df8a9c81fb662ab3"} Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.636637 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-7p5wk" Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.638457 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"761374dd-700e-42fa-81d4-d560f97643be","Type":"ContainerStarted","Data":"3010686321cfbdbdfa116a1e1351d3a92ecc31b80246b9425e166cecbb508cbb"} Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.640778 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3f18961c-c654-4f94-98a3-697644fdfda4","Type":"ContainerStarted","Data":"80b5d1a5f1649ecdbbf966aee951ba3b285f563dc0d4e51748d6d5e6cf62c5cd"} Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.667063 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=32.22985973 podStartE2EDuration="37.667040859s" podCreationTimestamp="2025-12-05 08:33:52 +0000 UTC" firstStartedPulling="2025-12-05 08:34:23.244842673 +0000 UTC m=+1127.733507295" lastFinishedPulling="2025-12-05 08:34:28.682023802 +0000 UTC m=+1133.170688424" observedRunningTime="2025-12-05 08:34:29.661275796 +0000 UTC m=+1134.149940438" watchObservedRunningTime="2025-12-05 08:34:29.667040859 +0000 UTC m=+1134.155705501" Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.683186 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-7p5wk" podStartSLOduration=29.396008488 podStartE2EDuration="34.68315893s" podCreationTimestamp="2025-12-05 08:33:55 +0000 UTC" firstStartedPulling="2025-12-05 08:34:23.273149776 +0000 UTC m=+1127.761814398" lastFinishedPulling="2025-12-05 08:34:28.560300218 +0000 UTC m=+1133.048964840" observedRunningTime="2025-12-05 08:34:29.677701514 +0000 UTC m=+1134.166366136" watchObservedRunningTime="2025-12-05 08:34:29.68315893 +0000 UTC m=+1134.171823572" Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.706484 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=40.259054706 podStartE2EDuration="40.706469689s" podCreationTimestamp="2025-12-05 08:33:49 +0000 UTC" firstStartedPulling="2025-12-05 08:34:23.100740604 +0000 UTC m=+1127.589405226" lastFinishedPulling="2025-12-05 08:34:23.548155577 +0000 UTC m=+1128.036820209" observedRunningTime="2025-12-05 08:34:29.701552537 +0000 UTC m=+1134.190217159" watchObservedRunningTime="2025-12-05 08:34:29.706469689 +0000 UTC m=+1134.195134301" Dec 05 08:34:29 crc kubenswrapper[4876]: I1205 08:34:29.719718 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=35.132262707 podStartE2EDuration="39.719700568s" podCreationTimestamp="2025-12-05 08:33:50 +0000 UTC" firstStartedPulling="2025-12-05 08:34:23.282088918 +0000 UTC m=+1127.770753540" lastFinishedPulling="2025-12-05 08:34:27.869526779 +0000 UTC m=+1132.358191401" observedRunningTime="2025-12-05 08:34:29.715031622 +0000 UTC m=+1134.203696244" watchObservedRunningTime="2025-12-05 08:34:29.719700568 +0000 UTC m=+1134.208365190" Dec 05 08:34:30 crc kubenswrapper[4876]: I1205 08:34:30.655865 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhwc7" event={"ID":"07839a94-e94b-4ff2-9eaf-9fb315c83046","Type":"ContainerStarted","Data":"867d2115f936a4f93a2e0a80a229e3f3e5f6e6a3769924587c04a5e1a46301ce"} Dec 05 08:34:30 crc kubenswrapper[4876]: I1205 08:34:30.657395 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:34:30 crc kubenswrapper[4876]: I1205 08:34:30.657424 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-fhwc7" event={"ID":"07839a94-e94b-4ff2-9eaf-9fb315c83046","Type":"ContainerStarted","Data":"529620b5db0b03acf36aadf2390c6323758956d1b9b133e4dcaefea680311f8b"} Dec 05 08:34:30 crc kubenswrapper[4876]: I1205 08:34:30.659187 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45b18655-19ec-43f4-a779-b8f72bc77aa2","Type":"ContainerStarted","Data":"f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369"} Dec 05 08:34:30 crc kubenswrapper[4876]: I1205 08:34:30.686207 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-fhwc7" podStartSLOduration=31.209878794 podStartE2EDuration="35.686184034s" podCreationTimestamp="2025-12-05 08:33:55 +0000 UTC" firstStartedPulling="2025-12-05 08:34:23.394196223 +0000 UTC m=+1127.882860845" lastFinishedPulling="2025-12-05 08:34:27.870501453 +0000 UTC m=+1132.359166085" observedRunningTime="2025-12-05 08:34:30.673348326 +0000 UTC m=+1135.162012958" watchObservedRunningTime="2025-12-05 08:34:30.686184034 +0000 UTC m=+1135.174848666" Dec 05 08:34:30 crc kubenswrapper[4876]: I1205 08:34:30.820600 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:34:30 crc kubenswrapper[4876]: I1205 08:34:30.848621 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 05 08:34:30 crc kubenswrapper[4876]: I1205 08:34:30.848664 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 05 08:34:33 crc kubenswrapper[4876]: I1205 08:34:33.685867 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9c3699c2-facd-4d82-92ec-4816b871833b","Type":"ContainerStarted","Data":"12054c33839f508d80b365eca5a66dbae4f6f3f26ff63c3543c8b1d513862967"} Dec 05 08:34:33 crc kubenswrapper[4876]: I1205 08:34:33.689278 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3f18961c-c654-4f94-98a3-697644fdfda4","Type":"ContainerStarted","Data":"c8ffcf43e5a569a678dc115857226ffc9c14fdfc5ecaa422ce43e91e47df593c"} Dec 05 08:34:33 crc kubenswrapper[4876]: I1205 08:34:33.705088 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=25.124283717 podStartE2EDuration="34.705073944s" podCreationTimestamp="2025-12-05 08:33:59 +0000 UTC" firstStartedPulling="2025-12-05 08:34:23.094357245 +0000 UTC m=+1127.583021867" lastFinishedPulling="2025-12-05 08:34:32.675147462 +0000 UTC m=+1137.163812094" observedRunningTime="2025-12-05 08:34:33.702179852 +0000 UTC m=+1138.190844464" watchObservedRunningTime="2025-12-05 08:34:33.705073944 +0000 UTC m=+1138.193738566" Dec 05 08:34:33 crc kubenswrapper[4876]: I1205 08:34:33.731048 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:33 crc kubenswrapper[4876]: I1205 08:34:33.731962 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=29.236105775 podStartE2EDuration="38.73189161s" podCreationTimestamp="2025-12-05 08:33:55 +0000 UTC" firstStartedPulling="2025-12-05 08:34:23.16217923 +0000 UTC m=+1127.650843852" lastFinishedPulling="2025-12-05 08:34:32.657965065 +0000 UTC m=+1137.146629687" observedRunningTime="2025-12-05 08:34:33.729031409 +0000 UTC m=+1138.217696031" watchObservedRunningTime="2025-12-05 08:34:33.73189161 +0000 UTC m=+1138.220556242" Dec 05 08:34:33 crc kubenswrapper[4876]: I1205 08:34:33.772846 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:34 crc kubenswrapper[4876]: I1205 08:34:34.701254 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:34 crc kubenswrapper[4876]: I1205 08:34:34.821519 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 05 08:34:34 crc kubenswrapper[4876]: I1205 08:34:34.952490 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.186764 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.188989 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x2c7b"] Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.201861 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-hxnnn"] Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.205195 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.206770 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.232773 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hxnnn"] Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.243794 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gx2x4"] Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.245443 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.248963 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.254406 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gx2x4"] Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.352608 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.356877 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/82469e38-4872-4d09-8a1e-2e635df62905-ovs-rundir\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.356943 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82469e38-4872-4d09-8a1e-2e635df62905-combined-ca-bundle\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.356975 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82469e38-4872-4d09-8a1e-2e635df62905-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.357097 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82469e38-4872-4d09-8a1e-2e635df62905-config\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.357171 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.357191 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/82469e38-4872-4d09-8a1e-2e635df62905-ovn-rundir\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.357223 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-config\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.357247 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.357311 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5cll\" (UniqueName: \"kubernetes.io/projected/82469e38-4872-4d09-8a1e-2e635df62905-kube-api-access-d5cll\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.357331 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk8kc\" (UniqueName: \"kubernetes.io/projected/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-kube-api-access-mk8kc\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.393556 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.458878 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/82469e38-4872-4d09-8a1e-2e635df62905-ovs-rundir\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.458955 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82469e38-4872-4d09-8a1e-2e635df62905-combined-ca-bundle\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.458986 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82469e38-4872-4d09-8a1e-2e635df62905-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459011 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82469e38-4872-4d09-8a1e-2e635df62905-config\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459037 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459114 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/82469e38-4872-4d09-8a1e-2e635df62905-ovn-rundir\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459134 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-config\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459151 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459179 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5cll\" (UniqueName: \"kubernetes.io/projected/82469e38-4872-4d09-8a1e-2e635df62905-kube-api-access-d5cll\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459193 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk8kc\" (UniqueName: \"kubernetes.io/projected/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-kube-api-access-mk8kc\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459255 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/82469e38-4872-4d09-8a1e-2e635df62905-ovn-rundir\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.459855 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82469e38-4872-4d09-8a1e-2e635df62905-config\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.460206 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.460245 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-config\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.460276 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/82469e38-4872-4d09-8a1e-2e635df62905-ovs-rundir\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.460283 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.464065 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82469e38-4872-4d09-8a1e-2e635df62905-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.467673 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82469e38-4872-4d09-8a1e-2e635df62905-combined-ca-bundle\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.475958 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5cll\" (UniqueName: \"kubernetes.io/projected/82469e38-4872-4d09-8a1e-2e635df62905-kube-api-access-d5cll\") pod \"ovn-controller-metrics-hxnnn\" (UID: \"82469e38-4872-4d09-8a1e-2e635df62905\") " pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.478016 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk8kc\" (UniqueName: \"kubernetes.io/projected/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-kube-api-access-mk8kc\") pod \"dnsmasq-dns-6bc7876d45-gx2x4\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.520558 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hxnnn" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.562481 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dtnkx"] Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.573153 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.670830 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-gs4d5"] Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.672359 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.675188 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.678355 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gs4d5"] Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.726332 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f","Type":"ContainerStarted","Data":"59bcde78d239d2ec5b2c1b6c42bfe979ce595826cff0ba2f07bdfb82de0dac70"} Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.728995 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" podUID="03da7075-fc4f-42e1-bc60-0acdf6fef427" containerName="init" containerID="cri-o://85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4" gracePeriod=10 Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.729197 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" event={"ID":"03da7075-fc4f-42e1-bc60-0acdf6fef427","Type":"ContainerStarted","Data":"85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4"} Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.729223 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.772420 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-dns-svc\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.772741 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.772771 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.772791 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvl49\" (UniqueName: \"kubernetes.io/projected/b95c35f4-164c-4466-954e-6005407caca5-kube-api-access-nvl49\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.772869 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-config\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.784582 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.857159 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.875350 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.875442 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvl49\" (UniqueName: \"kubernetes.io/projected/b95c35f4-164c-4466-954e-6005407caca5-kube-api-access-nvl49\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.879103 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.887126 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-config\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.887815 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-dns-svc\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.887890 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.887889 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.889383 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-config\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.889695 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.890409 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-dns-svc\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.896373 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvl49\" (UniqueName: \"kubernetes.io/projected/b95c35f4-164c-4466-954e-6005407caca5-kube-api-access-nvl49\") pod \"dnsmasq-dns-8554648995-gs4d5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:35 crc kubenswrapper[4876]: I1205 08:34:35.993571 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.167029 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.169142 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.172563 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.172939 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-fxtb7" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.173079 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.173177 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.176132 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.248887 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.294621 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-dns-svc\") pod \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.294697 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv9p2\" (UniqueName: \"kubernetes.io/projected/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-kube-api-access-wv9p2\") pod \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.294793 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-config\") pod \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\" (UID: \"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6\") " Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.294981 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-config\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.295026 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.295051 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.295095 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-scripts\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.295148 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh6n8\" (UniqueName: \"kubernetes.io/projected/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-kube-api-access-kh6n8\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.295171 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.295196 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.295929 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6e5de03-4ce5-4c61-8589-7cd403a5e7a6" (UID: "a6e5de03-4ce5-4c61-8589-7cd403a5e7a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.296812 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-config" (OuterVolumeSpecName: "config") pod "a6e5de03-4ce5-4c61-8589-7cd403a5e7a6" (UID: "a6e5de03-4ce5-4c61-8589-7cd403a5e7a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.300131 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-kube-api-access-wv9p2" (OuterVolumeSpecName: "kube-api-access-wv9p2") pod "a6e5de03-4ce5-4c61-8589-7cd403a5e7a6" (UID: "a6e5de03-4ce5-4c61-8589-7cd403a5e7a6"). InnerVolumeSpecName "kube-api-access-wv9p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396601 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh6n8\" (UniqueName: \"kubernetes.io/projected/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-kube-api-access-kh6n8\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396650 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396692 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396750 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-config\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396783 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396807 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396854 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-scripts\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396955 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396966 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.396975 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv9p2\" (UniqueName: \"kubernetes.io/projected/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6-kube-api-access-wv9p2\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.397926 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-scripts\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.398003 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.398428 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-config\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.402272 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.402311 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.420951 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh6n8\" (UniqueName: \"kubernetes.io/projected/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-kube-api-access-kh6n8\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.421161 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f6758b3-0cba-4d9b-b3ad-25149d206ff9-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f6758b3-0cba-4d9b-b3ad-25149d206ff9\") " pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.451254 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hxnnn"] Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.486588 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.513024 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gs4d5"] Dec 05 08:34:36 crc kubenswrapper[4876]: W1205 08:34:36.519259 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb95c35f4_164c_4466_954e_6005407caca5.slice/crio-299c622dbdd1187b8aa14718ff797abe2f099b09a90c7cf671b0c3a7b05764af WatchSource:0}: Error finding container 299c622dbdd1187b8aa14718ff797abe2f099b09a90c7cf671b0c3a7b05764af: Status 404 returned error can't find the container with id 299c622dbdd1187b8aa14718ff797abe2f099b09a90c7cf671b0c3a7b05764af Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.522743 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.555790 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gx2x4"] Dec 05 08:34:36 crc kubenswrapper[4876]: W1205 08:34:36.556179 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod599f38b0_ceda_4bb8_8bd0_01dd59fd03fe.slice/crio-59a1cd13058f69eb210f45f746078dce8cdfcfd103bd4ded12fca8f45cba3bcf WatchSource:0}: Error finding container 59a1cd13058f69eb210f45f746078dce8cdfcfd103bd4ded12fca8f45cba3bcf: Status 404 returned error can't find the container with id 59a1cd13058f69eb210f45f746078dce8cdfcfd103bd4ded12fca8f45cba3bcf Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.599046 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-config\") pod \"03da7075-fc4f-42e1-bc60-0acdf6fef427\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.599451 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-672ft\" (UniqueName: \"kubernetes.io/projected/03da7075-fc4f-42e1-bc60-0acdf6fef427-kube-api-access-672ft\") pod \"03da7075-fc4f-42e1-bc60-0acdf6fef427\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.599619 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-dns-svc\") pod \"03da7075-fc4f-42e1-bc60-0acdf6fef427\" (UID: \"03da7075-fc4f-42e1-bc60-0acdf6fef427\") " Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.603124 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03da7075-fc4f-42e1-bc60-0acdf6fef427-kube-api-access-672ft" (OuterVolumeSpecName: "kube-api-access-672ft") pod "03da7075-fc4f-42e1-bc60-0acdf6fef427" (UID: "03da7075-fc4f-42e1-bc60-0acdf6fef427"). InnerVolumeSpecName "kube-api-access-672ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.617535 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "03da7075-fc4f-42e1-bc60-0acdf6fef427" (UID: "03da7075-fc4f-42e1-bc60-0acdf6fef427"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.617704 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-config" (OuterVolumeSpecName: "config") pod "03da7075-fc4f-42e1-bc60-0acdf6fef427" (UID: "03da7075-fc4f-42e1-bc60-0acdf6fef427"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.702096 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.702135 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-672ft\" (UniqueName: \"kubernetes.io/projected/03da7075-fc4f-42e1-bc60-0acdf6fef427-kube-api-access-672ft\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.702147 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03da7075-fc4f-42e1-bc60-0acdf6fef427-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.741609 4876 generic.go:334] "Generic (PLEG): container finished" podID="03da7075-fc4f-42e1-bc60-0acdf6fef427" containerID="85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4" exitCode=0 Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.741681 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" event={"ID":"03da7075-fc4f-42e1-bc60-0acdf6fef427","Type":"ContainerDied","Data":"85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4"} Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.741711 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" event={"ID":"03da7075-fc4f-42e1-bc60-0acdf6fef427","Type":"ContainerDied","Data":"c3206172c9e527a4f9f9136cee66f8b152ce3bdc3d2d7a3429da15939e701b48"} Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.741732 4876 scope.go:117] "RemoveContainer" containerID="85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.741857 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-x2c7b" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.747368 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d47eff06-2977-4bcb-bc3b-6ad08399bbd7","Type":"ContainerStarted","Data":"9b047c0047451bb8bf87d9a9b851f5c530b1a37e275237fcef1056793b2a5c51"} Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.750483 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.750589 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dtnkx" event={"ID":"a6e5de03-4ce5-4c61-8589-7cd403a5e7a6","Type":"ContainerDied","Data":"64550113a8561e8fedef1ff37e0f739c6631ee31df88fe0db26bbc62f79baafa"} Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.757061 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" event={"ID":"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe","Type":"ContainerStarted","Data":"59a1cd13058f69eb210f45f746078dce8cdfcfd103bd4ded12fca8f45cba3bcf"} Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.758766 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gs4d5" event={"ID":"b95c35f4-164c-4466-954e-6005407caca5","Type":"ContainerStarted","Data":"299c622dbdd1187b8aa14718ff797abe2f099b09a90c7cf671b0c3a7b05764af"} Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.764364 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hxnnn" event={"ID":"82469e38-4872-4d09-8a1e-2e635df62905","Type":"ContainerStarted","Data":"4cb897adb72a96df697d21d407f6295a07430f66575f44855d272c56055dd9a3"} Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.782009 4876 scope.go:117] "RemoveContainer" containerID="85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4" Dec 05 08:34:36 crc kubenswrapper[4876]: E1205 08:34:36.792377 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4\": container with ID starting with 85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4 not found: ID does not exist" containerID="85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.792430 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4"} err="failed to get container status \"85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4\": rpc error: code = NotFound desc = could not find container \"85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4\": container with ID starting with 85f5778ada99d40708732b8b2ef0a1f80d228b4089524f4fa4da700d8e608ac4 not found: ID does not exist" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.875443 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dtnkx"] Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.889583 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-hxnnn" podStartSLOduration=1.8895694170000001 podStartE2EDuration="1.889569417s" podCreationTimestamp="2025-12-05 08:34:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:34:36.876933644 +0000 UTC m=+1141.365598266" watchObservedRunningTime="2025-12-05 08:34:36.889569417 +0000 UTC m=+1141.378234039" Dec 05 08:34:36 crc kubenswrapper[4876]: I1205 08:34:36.893971 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dtnkx"] Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.002325 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.050311 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x2c7b"] Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.059509 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-x2c7b"] Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.771573 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f6758b3-0cba-4d9b-b3ad-25149d206ff9","Type":"ContainerStarted","Data":"06ee6f6b058be7513ae443773fa10fc098e391ea4400504ab6ec94ae6c445abb"} Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.773890 4876 generic.go:334] "Generic (PLEG): container finished" podID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" containerID="87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343" exitCode=0 Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.773968 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" event={"ID":"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe","Type":"ContainerDied","Data":"87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343"} Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.778682 4876 generic.go:334] "Generic (PLEG): container finished" podID="b95c35f4-164c-4466-954e-6005407caca5" containerID="a04f5489ec8e7b5f673febde7b0768b66fbed680c5ea48cbb16b83851551a9fa" exitCode=0 Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.778750 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gs4d5" event={"ID":"b95c35f4-164c-4466-954e-6005407caca5","Type":"ContainerDied","Data":"a04f5489ec8e7b5f673febde7b0768b66fbed680c5ea48cbb16b83851551a9fa"} Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.781817 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hxnnn" event={"ID":"82469e38-4872-4d09-8a1e-2e635df62905","Type":"ContainerStarted","Data":"f4dd8e1f8419d7f2d87514a569357d60f8bb5999f17ef37e83fcc6116d530426"} Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.834021 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03da7075-fc4f-42e1-bc60-0acdf6fef427" path="/var/lib/kubelet/pods/03da7075-fc4f-42e1-bc60-0acdf6fef427/volumes" Dec 05 08:34:37 crc kubenswrapper[4876]: I1205 08:34:37.834637 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6e5de03-4ce5-4c61-8589-7cd403a5e7a6" path="/var/lib/kubelet/pods/a6e5de03-4ce5-4c61-8589-7cd403a5e7a6/volumes" Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.791279 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f6758b3-0cba-4d9b-b3ad-25149d206ff9","Type":"ContainerStarted","Data":"957a73fe021e083b886bf3cfbf2515996b2ac8cd0f28ecf3178d26cc9ba81332"} Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.791347 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f6758b3-0cba-4d9b-b3ad-25149d206ff9","Type":"ContainerStarted","Data":"0496d5b325665a43a89b7e5820388d3efc1fdb296b4b8b112de2879c50bc4e0a"} Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.791484 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.794924 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" event={"ID":"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe","Type":"ContainerStarted","Data":"6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd"} Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.795007 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.798740 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gs4d5" event={"ID":"b95c35f4-164c-4466-954e-6005407caca5","Type":"ContainerStarted","Data":"99383be0e4efcabad0cf22226748cb90878ac40c5f56b578a56202004959f124"} Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.816609 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.699559958 podStartE2EDuration="2.816585034s" podCreationTimestamp="2025-12-05 08:34:36 +0000 UTC" firstStartedPulling="2025-12-05 08:34:37.004689787 +0000 UTC m=+1141.493354409" lastFinishedPulling="2025-12-05 08:34:38.121714863 +0000 UTC m=+1142.610379485" observedRunningTime="2025-12-05 08:34:38.811415805 +0000 UTC m=+1143.300080437" watchObservedRunningTime="2025-12-05 08:34:38.816585034 +0000 UTC m=+1143.305249676" Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.850947 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-gs4d5" podStartSLOduration=3.850884306 podStartE2EDuration="3.850884306s" podCreationTimestamp="2025-12-05 08:34:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:34:38.840004475 +0000 UTC m=+1143.328669127" watchObservedRunningTime="2025-12-05 08:34:38.850884306 +0000 UTC m=+1143.339548968" Dec 05 08:34:38 crc kubenswrapper[4876]: I1205 08:34:38.877595 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" podStartSLOduration=3.877570839 podStartE2EDuration="3.877570839s" podCreationTimestamp="2025-12-05 08:34:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:34:38.872341399 +0000 UTC m=+1143.361006061" watchObservedRunningTime="2025-12-05 08:34:38.877570839 +0000 UTC m=+1143.366235501" Dec 05 08:34:39 crc kubenswrapper[4876]: I1205 08:34:39.809360 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:40 crc kubenswrapper[4876]: I1205 08:34:40.822630 4876 generic.go:334] "Generic (PLEG): container finished" podID="d47eff06-2977-4bcb-bc3b-6ad08399bbd7" containerID="9b047c0047451bb8bf87d9a9b851f5c530b1a37e275237fcef1056793b2a5c51" exitCode=0 Dec 05 08:34:40 crc kubenswrapper[4876]: I1205 08:34:40.822722 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d47eff06-2977-4bcb-bc3b-6ad08399bbd7","Type":"ContainerDied","Data":"9b047c0047451bb8bf87d9a9b851f5c530b1a37e275237fcef1056793b2a5c51"} Dec 05 08:34:41 crc kubenswrapper[4876]: I1205 08:34:41.843527 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d47eff06-2977-4bcb-bc3b-6ad08399bbd7","Type":"ContainerStarted","Data":"1ca5ab7c9a815edce55b8dacf82608c52c65b57c45b26ef53683f2935dfe75bc"} Dec 05 08:34:41 crc kubenswrapper[4876]: I1205 08:34:41.886561 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371981.96825 podStartE2EDuration="54.88652567s" podCreationTimestamp="2025-12-05 08:33:47 +0000 UTC" firstStartedPulling="2025-12-05 08:33:49.931924882 +0000 UTC m=+1094.420589494" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:34:41.87567743 +0000 UTC m=+1146.364342092" watchObservedRunningTime="2025-12-05 08:34:41.88652567 +0000 UTC m=+1146.375190322" Dec 05 08:34:42 crc kubenswrapper[4876]: I1205 08:34:42.780376 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gx2x4"] Dec 05 08:34:42 crc kubenswrapper[4876]: I1205 08:34:42.780917 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" podUID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" containerName="dnsmasq-dns" containerID="cri-o://6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd" gracePeriod=10 Dec 05 08:34:42 crc kubenswrapper[4876]: I1205 08:34:42.808993 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 08:34:42 crc kubenswrapper[4876]: I1205 08:34:42.827107 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gvpkj"] Dec 05 08:34:42 crc kubenswrapper[4876]: E1205 08:34:42.827435 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03da7075-fc4f-42e1-bc60-0acdf6fef427" containerName="init" Dec 05 08:34:42 crc kubenswrapper[4876]: I1205 08:34:42.827453 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="03da7075-fc4f-42e1-bc60-0acdf6fef427" containerName="init" Dec 05 08:34:42 crc kubenswrapper[4876]: I1205 08:34:42.827626 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="03da7075-fc4f-42e1-bc60-0acdf6fef427" containerName="init" Dec 05 08:34:42 crc kubenswrapper[4876]: I1205 08:34:42.844724 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:42 crc kubenswrapper[4876]: I1205 08:34:42.864626 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gvpkj"] Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.005065 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-config\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.005145 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.005173 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.005497 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxbx4\" (UniqueName: \"kubernetes.io/projected/c3dadbab-64b6-44e6-b07c-bc5372ffb775-kube-api-access-gxbx4\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.005547 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.107217 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-config\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.107577 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.107602 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.107706 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxbx4\" (UniqueName: \"kubernetes.io/projected/c3dadbab-64b6-44e6-b07c-bc5372ffb775-kube-api-access-gxbx4\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.107735 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.108093 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-config\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.108650 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.108659 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.109220 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.144381 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxbx4\" (UniqueName: \"kubernetes.io/projected/c3dadbab-64b6-44e6-b07c-bc5372ffb775-kube-api-access-gxbx4\") pod \"dnsmasq-dns-b8fbc5445-gvpkj\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.199514 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.250076 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.309667 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk8kc\" (UniqueName: \"kubernetes.io/projected/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-kube-api-access-mk8kc\") pod \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.309734 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-config\") pod \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.309871 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-dns-svc\") pod \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.309977 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-ovsdbserver-sb\") pod \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\" (UID: \"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe\") " Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.316629 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-kube-api-access-mk8kc" (OuterVolumeSpecName: "kube-api-access-mk8kc") pod "599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" (UID: "599f38b0-ceda-4bb8-8bd0-01dd59fd03fe"). InnerVolumeSpecName "kube-api-access-mk8kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.356498 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-config" (OuterVolumeSpecName: "config") pod "599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" (UID: "599f38b0-ceda-4bb8-8bd0-01dd59fd03fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.371736 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" (UID: "599f38b0-ceda-4bb8-8bd0-01dd59fd03fe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.377854 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" (UID: "599f38b0-ceda-4bb8-8bd0-01dd59fd03fe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.412943 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.412981 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.412992 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.413007 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk8kc\" (UniqueName: \"kubernetes.io/projected/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe-kube-api-access-mk8kc\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.690812 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gvpkj"] Dec 05 08:34:43 crc kubenswrapper[4876]: W1205 08:34:43.694811 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3dadbab_64b6_44e6_b07c_bc5372ffb775.slice/crio-e1bf612f2cdc05e67bbbc71bfb3c70c231487c295967e6481844a853eb93658f WatchSource:0}: Error finding container e1bf612f2cdc05e67bbbc71bfb3c70c231487c295967e6481844a853eb93658f: Status 404 returned error can't find the container with id e1bf612f2cdc05e67bbbc71bfb3c70c231487c295967e6481844a853eb93658f Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.864676 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" event={"ID":"c3dadbab-64b6-44e6-b07c-bc5372ffb775","Type":"ContainerStarted","Data":"e1bf612f2cdc05e67bbbc71bfb3c70c231487c295967e6481844a853eb93658f"} Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.867155 4876 generic.go:334] "Generic (PLEG): container finished" podID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" containerID="6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd" exitCode=0 Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.867200 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" event={"ID":"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe","Type":"ContainerDied","Data":"6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd"} Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.867224 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.867257 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-gx2x4" event={"ID":"599f38b0-ceda-4bb8-8bd0-01dd59fd03fe","Type":"ContainerDied","Data":"59a1cd13058f69eb210f45f746078dce8cdfcfd103bd4ded12fca8f45cba3bcf"} Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.867279 4876 scope.go:117] "RemoveContainer" containerID="6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.893825 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gx2x4"] Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.899116 4876 scope.go:117] "RemoveContainer" containerID="87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.910720 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-gx2x4"] Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.920970 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 05 08:34:43 crc kubenswrapper[4876]: E1205 08:34:43.921400 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" containerName="dnsmasq-dns" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.921425 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" containerName="dnsmasq-dns" Dec 05 08:34:43 crc kubenswrapper[4876]: E1205 08:34:43.921446 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" containerName="init" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.921454 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" containerName="init" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.921823 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" containerName="dnsmasq-dns" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.928201 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.930765 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-65bhr" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.931086 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.931235 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.931443 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.934033 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.975807 4876 scope.go:117] "RemoveContainer" containerID="6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd" Dec 05 08:34:43 crc kubenswrapper[4876]: E1205 08:34:43.976381 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd\": container with ID starting with 6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd not found: ID does not exist" containerID="6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.976412 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd"} err="failed to get container status \"6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd\": rpc error: code = NotFound desc = could not find container \"6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd\": container with ID starting with 6154b0b9b806d653a7b7fef4258e63d7cff8aca2853ea3e741e785cfe3c8f1dd not found: ID does not exist" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.976555 4876 scope.go:117] "RemoveContainer" containerID="87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343" Dec 05 08:34:43 crc kubenswrapper[4876]: E1205 08:34:43.977641 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343\": container with ID starting with 87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343 not found: ID does not exist" containerID="87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343" Dec 05 08:34:43 crc kubenswrapper[4876]: I1205 08:34:43.977682 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343"} err="failed to get container status \"87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343\": rpc error: code = NotFound desc = could not find container \"87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343\": container with ID starting with 87e036a05b5b6ba80f57395adc90b6d5096325ccba909e43ca4c68b1d74ea343 not found: ID does not exist" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.020352 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxdw9\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-kube-api-access-kxdw9\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.020417 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/71735818-3684-481f-bbb3-72aff2d50428-cache\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.020580 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/71735818-3684-481f-bbb3-72aff2d50428-lock\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.020678 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.020782 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.122662 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxdw9\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-kube-api-access-kxdw9\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.122720 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/71735818-3684-481f-bbb3-72aff2d50428-cache\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.122754 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/71735818-3684-481f-bbb3-72aff2d50428-lock\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.122790 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.122838 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.123126 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: E1205 08:34:44.123207 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 08:34:44 crc kubenswrapper[4876]: E1205 08:34:44.123239 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 08:34:44 crc kubenswrapper[4876]: E1205 08:34:44.123298 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift podName:71735818-3684-481f-bbb3-72aff2d50428 nodeName:}" failed. No retries permitted until 2025-12-05 08:34:44.62327308 +0000 UTC m=+1149.111937702 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift") pod "swift-storage-0" (UID: "71735818-3684-481f-bbb3-72aff2d50428") : configmap "swift-ring-files" not found Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.123345 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/71735818-3684-481f-bbb3-72aff2d50428-lock\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.123625 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/71735818-3684-481f-bbb3-72aff2d50428-cache\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.147099 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxdw9\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-kube-api-access-kxdw9\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.151267 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.373358 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-hzjsf"] Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.374687 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.377970 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.379326 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.381477 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.388327 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hzjsf"] Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.529825 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-scripts\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.530389 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzgb8\" (UniqueName: \"kubernetes.io/projected/d41c7aad-b564-4c13-9f60-2d1fe53861a8-kube-api-access-gzgb8\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.530510 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-swiftconf\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.530611 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d41c7aad-b564-4c13-9f60-2d1fe53861a8-etc-swift\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.530689 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-dispersionconf\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.530768 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-combined-ca-bundle\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.530881 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-ring-data-devices\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.632545 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-scripts\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.633511 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-scripts\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.633608 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.633696 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzgb8\" (UniqueName: \"kubernetes.io/projected/d41c7aad-b564-4c13-9f60-2d1fe53861a8-kube-api-access-gzgb8\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: E1205 08:34:44.633833 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 08:34:44 crc kubenswrapper[4876]: E1205 08:34:44.633864 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 08:34:44 crc kubenswrapper[4876]: E1205 08:34:44.633933 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift podName:71735818-3684-481f-bbb3-72aff2d50428 nodeName:}" failed. No retries permitted until 2025-12-05 08:34:45.633913804 +0000 UTC m=+1150.122578426 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift") pod "swift-storage-0" (UID: "71735818-3684-481f-bbb3-72aff2d50428") : configmap "swift-ring-files" not found Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.633780 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-swiftconf\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.634200 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d41c7aad-b564-4c13-9f60-2d1fe53861a8-etc-swift\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.634287 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-dispersionconf\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.634623 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d41c7aad-b564-4c13-9f60-2d1fe53861a8-etc-swift\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.634321 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-combined-ca-bundle\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.635038 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-ring-data-devices\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.635501 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-ring-data-devices\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.637722 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-dispersionconf\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.637809 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-combined-ca-bundle\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.646559 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-swiftconf\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.647456 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzgb8\" (UniqueName: \"kubernetes.io/projected/d41c7aad-b564-4c13-9f60-2d1fe53861a8-kube-api-access-gzgb8\") pod \"swift-ring-rebalance-hzjsf\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.700803 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.880718 4876 generic.go:334] "Generic (PLEG): container finished" podID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerID="5636413e75e741732c3d3120266d0e141d7608ff5bb6ae372a17715d70b24e75" exitCode=0 Dec 05 08:34:44 crc kubenswrapper[4876]: I1205 08:34:44.880768 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" event={"ID":"c3dadbab-64b6-44e6-b07c-bc5372ffb775","Type":"ContainerDied","Data":"5636413e75e741732c3d3120266d0e141d7608ff5bb6ae372a17715d70b24e75"} Dec 05 08:34:45 crc kubenswrapper[4876]: I1205 08:34:45.147035 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hzjsf"] Dec 05 08:34:45 crc kubenswrapper[4876]: I1205 08:34:45.665250 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:45 crc kubenswrapper[4876]: E1205 08:34:45.665457 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 08:34:45 crc kubenswrapper[4876]: E1205 08:34:45.665484 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 08:34:45 crc kubenswrapper[4876]: E1205 08:34:45.665561 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift podName:71735818-3684-481f-bbb3-72aff2d50428 nodeName:}" failed. No retries permitted until 2025-12-05 08:34:47.665539849 +0000 UTC m=+1152.154204471 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift") pod "swift-storage-0" (UID: "71735818-3684-481f-bbb3-72aff2d50428") : configmap "swift-ring-files" not found Dec 05 08:34:45 crc kubenswrapper[4876]: I1205 08:34:45.835183 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="599f38b0-ceda-4bb8-8bd0-01dd59fd03fe" path="/var/lib/kubelet/pods/599f38b0-ceda-4bb8-8bd0-01dd59fd03fe/volumes" Dec 05 08:34:45 crc kubenswrapper[4876]: I1205 08:34:45.889150 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hzjsf" event={"ID":"d41c7aad-b564-4c13-9f60-2d1fe53861a8","Type":"ContainerStarted","Data":"a816b20455147ccd217afa20c75379d002dd7119516aeb464157a934d5648549"} Dec 05 08:34:45 crc kubenswrapper[4876]: I1205 08:34:45.995086 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:47 crc kubenswrapper[4876]: I1205 08:34:47.696973 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:47 crc kubenswrapper[4876]: E1205 08:34:47.697115 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 08:34:47 crc kubenswrapper[4876]: E1205 08:34:47.697167 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 08:34:47 crc kubenswrapper[4876]: E1205 08:34:47.697230 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift podName:71735818-3684-481f-bbb3-72aff2d50428 nodeName:}" failed. No retries permitted until 2025-12-05 08:34:51.697211506 +0000 UTC m=+1156.185876128 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift") pod "swift-storage-0" (UID: "71735818-3684-481f-bbb3-72aff2d50428") : configmap "swift-ring-files" not found Dec 05 08:34:47 crc kubenswrapper[4876]: I1205 08:34:47.906526 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" event={"ID":"c3dadbab-64b6-44e6-b07c-bc5372ffb775","Type":"ContainerStarted","Data":"3a63fd7d71efca11be3d44d274811be02f2758a0769977e828e60ba2761b9b43"} Dec 05 08:34:47 crc kubenswrapper[4876]: I1205 08:34:47.906813 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:49 crc kubenswrapper[4876]: I1205 08:34:49.282928 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 05 08:34:49 crc kubenswrapper[4876]: I1205 08:34:49.282988 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 05 08:34:49 crc kubenswrapper[4876]: I1205 08:34:49.356518 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 05 08:34:49 crc kubenswrapper[4876]: I1205 08:34:49.374356 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" podStartSLOduration=7.374339705 podStartE2EDuration="7.374339705s" podCreationTimestamp="2025-12-05 08:34:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:34:47.926491971 +0000 UTC m=+1152.415156593" watchObservedRunningTime="2025-12-05 08:34:49.374339705 +0000 UTC m=+1153.863004327" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.024573 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.374362 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-348b-account-create-update-5dmk7"] Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.375738 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.380261 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.386650 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-348b-account-create-update-5dmk7"] Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.394335 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bwggg"] Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.400512 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.404491 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bwggg"] Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.446966 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b76471-1845-41af-aecd-ea80bf5ba97d-operator-scripts\") pod \"keystone-db-create-bwggg\" (UID: \"b8b76471-1845-41af-aecd-ea80bf5ba97d\") " pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.447017 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm872\" (UniqueName: \"kubernetes.io/projected/b8b76471-1845-41af-aecd-ea80bf5ba97d-kube-api-access-pm872\") pod \"keystone-db-create-bwggg\" (UID: \"b8b76471-1845-41af-aecd-ea80bf5ba97d\") " pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.447105 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbl6m\" (UniqueName: \"kubernetes.io/projected/2ee916a3-9c6b-4102-84b1-f921c334f152-kube-api-access-nbl6m\") pod \"keystone-348b-account-create-update-5dmk7\" (UID: \"2ee916a3-9c6b-4102-84b1-f921c334f152\") " pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.447262 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ee916a3-9c6b-4102-84b1-f921c334f152-operator-scripts\") pod \"keystone-348b-account-create-update-5dmk7\" (UID: \"2ee916a3-9c6b-4102-84b1-f921c334f152\") " pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.549220 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm872\" (UniqueName: \"kubernetes.io/projected/b8b76471-1845-41af-aecd-ea80bf5ba97d-kube-api-access-pm872\") pod \"keystone-db-create-bwggg\" (UID: \"b8b76471-1845-41af-aecd-ea80bf5ba97d\") " pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.549492 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbl6m\" (UniqueName: \"kubernetes.io/projected/2ee916a3-9c6b-4102-84b1-f921c334f152-kube-api-access-nbl6m\") pod \"keystone-348b-account-create-update-5dmk7\" (UID: \"2ee916a3-9c6b-4102-84b1-f921c334f152\") " pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.549591 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ee916a3-9c6b-4102-84b1-f921c334f152-operator-scripts\") pod \"keystone-348b-account-create-update-5dmk7\" (UID: \"2ee916a3-9c6b-4102-84b1-f921c334f152\") " pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.549631 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b76471-1845-41af-aecd-ea80bf5ba97d-operator-scripts\") pod \"keystone-db-create-bwggg\" (UID: \"b8b76471-1845-41af-aecd-ea80bf5ba97d\") " pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.550614 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b76471-1845-41af-aecd-ea80bf5ba97d-operator-scripts\") pod \"keystone-db-create-bwggg\" (UID: \"b8b76471-1845-41af-aecd-ea80bf5ba97d\") " pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.551166 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ee916a3-9c6b-4102-84b1-f921c334f152-operator-scripts\") pod \"keystone-348b-account-create-update-5dmk7\" (UID: \"2ee916a3-9c6b-4102-84b1-f921c334f152\") " pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.573148 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbl6m\" (UniqueName: \"kubernetes.io/projected/2ee916a3-9c6b-4102-84b1-f921c334f152-kube-api-access-nbl6m\") pod \"keystone-348b-account-create-update-5dmk7\" (UID: \"2ee916a3-9c6b-4102-84b1-f921c334f152\") " pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.573356 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm872\" (UniqueName: \"kubernetes.io/projected/b8b76471-1845-41af-aecd-ea80bf5ba97d-kube-api-access-pm872\") pod \"keystone-db-create-bwggg\" (UID: \"b8b76471-1845-41af-aecd-ea80bf5ba97d\") " pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.697200 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.697963 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-w4xpx"] Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.699125 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.711812 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-w4xpx"] Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.725381 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.752376 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-operator-scripts\") pod \"placement-db-create-w4xpx\" (UID: \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\") " pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.752453 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwhst\" (UniqueName: \"kubernetes.io/projected/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-kube-api-access-dwhst\") pod \"placement-db-create-w4xpx\" (UID: \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\") " pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.830962 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-96e4-account-create-update-6fzmf"] Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.832017 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.837259 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.843967 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-96e4-account-create-update-6fzmf"] Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.855338 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-operator-scripts\") pod \"placement-db-create-w4xpx\" (UID: \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\") " pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.855420 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwhst\" (UniqueName: \"kubernetes.io/projected/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-kube-api-access-dwhst\") pod \"placement-db-create-w4xpx\" (UID: \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\") " pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.856969 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-operator-scripts\") pod \"placement-db-create-w4xpx\" (UID: \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\") " pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.883569 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwhst\" (UniqueName: \"kubernetes.io/projected/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-kube-api-access-dwhst\") pod \"placement-db-create-w4xpx\" (UID: \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\") " pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.945872 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hzjsf" event={"ID":"d41c7aad-b564-4c13-9f60-2d1fe53861a8","Type":"ContainerStarted","Data":"25f444d0078306eb58644ca66e8c564a9c14c9c8d23d12c8acc2f46026721a6f"} Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.963173 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbvt5\" (UniqueName: \"kubernetes.io/projected/90c4575e-5eb6-4c24-8662-9c2c4863bc08-kube-api-access-bbvt5\") pod \"placement-96e4-account-create-update-6fzmf\" (UID: \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\") " pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.963233 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c4575e-5eb6-4c24-8662-9c2c4863bc08-operator-scripts\") pod \"placement-96e4-account-create-update-6fzmf\" (UID: \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\") " pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:50 crc kubenswrapper[4876]: I1205 08:34:50.970221 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-hzjsf" podStartSLOduration=2.370618943 podStartE2EDuration="6.970205186s" podCreationTimestamp="2025-12-05 08:34:44 +0000 UTC" firstStartedPulling="2025-12-05 08:34:45.141613715 +0000 UTC m=+1149.630278337" lastFinishedPulling="2025-12-05 08:34:49.741199958 +0000 UTC m=+1154.229864580" observedRunningTime="2025-12-05 08:34:50.967441468 +0000 UTC m=+1155.456106090" watchObservedRunningTime="2025-12-05 08:34:50.970205186 +0000 UTC m=+1155.458869808" Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.065499 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbvt5\" (UniqueName: \"kubernetes.io/projected/90c4575e-5eb6-4c24-8662-9c2c4863bc08-kube-api-access-bbvt5\") pod \"placement-96e4-account-create-update-6fzmf\" (UID: \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\") " pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.065584 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c4575e-5eb6-4c24-8662-9c2c4863bc08-operator-scripts\") pod \"placement-96e4-account-create-update-6fzmf\" (UID: \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\") " pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.067239 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c4575e-5eb6-4c24-8662-9c2c4863bc08-operator-scripts\") pod \"placement-96e4-account-create-update-6fzmf\" (UID: \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\") " pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.091138 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbvt5\" (UniqueName: \"kubernetes.io/projected/90c4575e-5eb6-4c24-8662-9c2c4863bc08-kube-api-access-bbvt5\") pod \"placement-96e4-account-create-update-6fzmf\" (UID: \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\") " pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.143556 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.155690 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.204975 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bwggg"] Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.306358 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-348b-account-create-update-5dmk7"] Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.577051 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-96e4-account-create-update-6fzmf"] Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.664629 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.707888 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:51 crc kubenswrapper[4876]: E1205 08:34:51.708029 4876 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 08:34:51 crc kubenswrapper[4876]: E1205 08:34:51.708054 4876 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 08:34:51 crc kubenswrapper[4876]: E1205 08:34:51.708111 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift podName:71735818-3684-481f-bbb3-72aff2d50428 nodeName:}" failed. No retries permitted until 2025-12-05 08:34:59.708093935 +0000 UTC m=+1164.196758557 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift") pod "swift-storage-0" (UID: "71735818-3684-481f-bbb3-72aff2d50428") : configmap "swift-ring-files" not found Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.900308 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-w4xpx"] Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.956745 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w4xpx" event={"ID":"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee","Type":"ContainerStarted","Data":"96e5d564ec6ac177b9ff81191ae02270d31e7c94affe78f818c3d10131e67b41"} Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.960556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-348b-account-create-update-5dmk7" event={"ID":"2ee916a3-9c6b-4102-84b1-f921c334f152","Type":"ContainerStarted","Data":"caa5713e485468b46bc0d0c39ad0747d52321484011832333e7b6ba42feedbec"} Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.961514 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bwggg" event={"ID":"b8b76471-1845-41af-aecd-ea80bf5ba97d","Type":"ContainerStarted","Data":"e01fa549d1ae1b66baf5762c1d6b06d7fcfab3035ec4b84ed994c5cec533779f"} Dec 05 08:34:51 crc kubenswrapper[4876]: I1205 08:34:51.963075 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96e4-account-create-update-6fzmf" event={"ID":"90c4575e-5eb6-4c24-8662-9c2c4863bc08","Type":"ContainerStarted","Data":"389f8010673b80b187f188cf0b94fb583f9b8377ca54dafb3e77df2f6f8e34d8"} Dec 05 08:34:52 crc kubenswrapper[4876]: I1205 08:34:52.969957 4876 generic.go:334] "Generic (PLEG): container finished" podID="90c4575e-5eb6-4c24-8662-9c2c4863bc08" containerID="70660ff27709f826aeef1b81f8f9df35fdcf1e65df3f8281a0a8ef2695ecfb58" exitCode=0 Dec 05 08:34:52 crc kubenswrapper[4876]: I1205 08:34:52.970003 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96e4-account-create-update-6fzmf" event={"ID":"90c4575e-5eb6-4c24-8662-9c2c4863bc08","Type":"ContainerDied","Data":"70660ff27709f826aeef1b81f8f9df35fdcf1e65df3f8281a0a8ef2695ecfb58"} Dec 05 08:34:52 crc kubenswrapper[4876]: I1205 08:34:52.972843 4876 generic.go:334] "Generic (PLEG): container finished" podID="f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee" containerID="ddbc9e75a03fa7ff0cec4ff0c2d5ee17066f3aa0201e614f12a1dc1f18d4cd61" exitCode=0 Dec 05 08:34:52 crc kubenswrapper[4876]: I1205 08:34:52.972872 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w4xpx" event={"ID":"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee","Type":"ContainerDied","Data":"ddbc9e75a03fa7ff0cec4ff0c2d5ee17066f3aa0201e614f12a1dc1f18d4cd61"} Dec 05 08:34:52 crc kubenswrapper[4876]: I1205 08:34:52.974358 4876 generic.go:334] "Generic (PLEG): container finished" podID="2ee916a3-9c6b-4102-84b1-f921c334f152" containerID="57777bc0ed58a1d0d7d50bdd58c5c6f106344fcfa8a341227b49ee4cdab7ba55" exitCode=0 Dec 05 08:34:52 crc kubenswrapper[4876]: I1205 08:34:52.974406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-348b-account-create-update-5dmk7" event={"ID":"2ee916a3-9c6b-4102-84b1-f921c334f152","Type":"ContainerDied","Data":"57777bc0ed58a1d0d7d50bdd58c5c6f106344fcfa8a341227b49ee4cdab7ba55"} Dec 05 08:34:52 crc kubenswrapper[4876]: I1205 08:34:52.976317 4876 generic.go:334] "Generic (PLEG): container finished" podID="b8b76471-1845-41af-aecd-ea80bf5ba97d" containerID="773631e225f32fc1597c11685caa78682e0a9bee7833e77607fe9a4583a06da9" exitCode=0 Dec 05 08:34:52 crc kubenswrapper[4876]: I1205 08:34:52.976362 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bwggg" event={"ID":"b8b76471-1845-41af-aecd-ea80bf5ba97d","Type":"ContainerDied","Data":"773631e225f32fc1597c11685caa78682e0a9bee7833e77607fe9a4583a06da9"} Dec 05 08:34:53 crc kubenswrapper[4876]: I1205 08:34:53.252057 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:34:53 crc kubenswrapper[4876]: I1205 08:34:53.317258 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gs4d5"] Dec 05 08:34:53 crc kubenswrapper[4876]: I1205 08:34:53.317594 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-gs4d5" podUID="b95c35f4-164c-4466-954e-6005407caca5" containerName="dnsmasq-dns" containerID="cri-o://99383be0e4efcabad0cf22226748cb90878ac40c5f56b578a56202004959f124" gracePeriod=10 Dec 05 08:34:53 crc kubenswrapper[4876]: I1205 08:34:53.984541 4876 generic.go:334] "Generic (PLEG): container finished" podID="b95c35f4-164c-4466-954e-6005407caca5" containerID="99383be0e4efcabad0cf22226748cb90878ac40c5f56b578a56202004959f124" exitCode=0 Dec 05 08:34:53 crc kubenswrapper[4876]: I1205 08:34:53.984618 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gs4d5" event={"ID":"b95c35f4-164c-4466-954e-6005407caca5","Type":"ContainerDied","Data":"99383be0e4efcabad0cf22226748cb90878ac40c5f56b578a56202004959f124"} Dec 05 08:34:53 crc kubenswrapper[4876]: I1205 08:34:53.985096 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gs4d5" event={"ID":"b95c35f4-164c-4466-954e-6005407caca5","Type":"ContainerDied","Data":"299c622dbdd1187b8aa14718ff797abe2f099b09a90c7cf671b0c3a7b05764af"} Dec 05 08:34:53 crc kubenswrapper[4876]: I1205 08:34:53.985116 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="299c622dbdd1187b8aa14718ff797abe2f099b09a90c7cf671b0c3a7b05764af" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.036738 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.075710 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-sb\") pod \"b95c35f4-164c-4466-954e-6005407caca5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.075820 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvl49\" (UniqueName: \"kubernetes.io/projected/b95c35f4-164c-4466-954e-6005407caca5-kube-api-access-nvl49\") pod \"b95c35f4-164c-4466-954e-6005407caca5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.075934 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-config\") pod \"b95c35f4-164c-4466-954e-6005407caca5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.075989 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-nb\") pod \"b95c35f4-164c-4466-954e-6005407caca5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.076043 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-dns-svc\") pod \"b95c35f4-164c-4466-954e-6005407caca5\" (UID: \"b95c35f4-164c-4466-954e-6005407caca5\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.170731 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b95c35f4-164c-4466-954e-6005407caca5-kube-api-access-nvl49" (OuterVolumeSpecName: "kube-api-access-nvl49") pod "b95c35f4-164c-4466-954e-6005407caca5" (UID: "b95c35f4-164c-4466-954e-6005407caca5"). InnerVolumeSpecName "kube-api-access-nvl49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.178338 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvl49\" (UniqueName: \"kubernetes.io/projected/b95c35f4-164c-4466-954e-6005407caca5-kube-api-access-nvl49\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.195400 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b95c35f4-164c-4466-954e-6005407caca5" (UID: "b95c35f4-164c-4466-954e-6005407caca5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.195744 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b95c35f4-164c-4466-954e-6005407caca5" (UID: "b95c35f4-164c-4466-954e-6005407caca5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.208675 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-config" (OuterVolumeSpecName: "config") pod "b95c35f4-164c-4466-954e-6005407caca5" (UID: "b95c35f4-164c-4466-954e-6005407caca5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.249469 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b95c35f4-164c-4466-954e-6005407caca5" (UID: "b95c35f4-164c-4466-954e-6005407caca5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.279686 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.279722 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.279734 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.279742 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b95c35f4-164c-4466-954e-6005407caca5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.516208 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.681181 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.685173 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbl6m\" (UniqueName: \"kubernetes.io/projected/2ee916a3-9c6b-4102-84b1-f921c334f152-kube-api-access-nbl6m\") pod \"2ee916a3-9c6b-4102-84b1-f921c334f152\" (UID: \"2ee916a3-9c6b-4102-84b1-f921c334f152\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.685202 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ee916a3-9c6b-4102-84b1-f921c334f152-operator-scripts\") pod \"2ee916a3-9c6b-4102-84b1-f921c334f152\" (UID: \"2ee916a3-9c6b-4102-84b1-f921c334f152\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.686137 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee916a3-9c6b-4102-84b1-f921c334f152-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2ee916a3-9c6b-4102-84b1-f921c334f152" (UID: "2ee916a3-9c6b-4102-84b1-f921c334f152"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.687464 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.690563 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee916a3-9c6b-4102-84b1-f921c334f152-kube-api-access-nbl6m" (OuterVolumeSpecName: "kube-api-access-nbl6m") pod "2ee916a3-9c6b-4102-84b1-f921c334f152" (UID: "2ee916a3-9c6b-4102-84b1-f921c334f152"). InnerVolumeSpecName "kube-api-access-nbl6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.691716 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.786778 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwhst\" (UniqueName: \"kubernetes.io/projected/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-kube-api-access-dwhst\") pod \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\" (UID: \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.787169 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-operator-scripts\") pod \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\" (UID: \"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.787573 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbl6m\" (UniqueName: \"kubernetes.io/projected/2ee916a3-9c6b-4102-84b1-f921c334f152-kube-api-access-nbl6m\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.787596 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ee916a3-9c6b-4102-84b1-f921c334f152-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.788069 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee" (UID: "f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.790622 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-kube-api-access-dwhst" (OuterVolumeSpecName: "kube-api-access-dwhst") pod "f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee" (UID: "f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee"). InnerVolumeSpecName "kube-api-access-dwhst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.888693 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbvt5\" (UniqueName: \"kubernetes.io/projected/90c4575e-5eb6-4c24-8662-9c2c4863bc08-kube-api-access-bbvt5\") pod \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\" (UID: \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.888810 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c4575e-5eb6-4c24-8662-9c2c4863bc08-operator-scripts\") pod \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\" (UID: \"90c4575e-5eb6-4c24-8662-9c2c4863bc08\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.888985 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm872\" (UniqueName: \"kubernetes.io/projected/b8b76471-1845-41af-aecd-ea80bf5ba97d-kube-api-access-pm872\") pod \"b8b76471-1845-41af-aecd-ea80bf5ba97d\" (UID: \"b8b76471-1845-41af-aecd-ea80bf5ba97d\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.889089 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b76471-1845-41af-aecd-ea80bf5ba97d-operator-scripts\") pod \"b8b76471-1845-41af-aecd-ea80bf5ba97d\" (UID: \"b8b76471-1845-41af-aecd-ea80bf5ba97d\") " Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.889538 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwhst\" (UniqueName: \"kubernetes.io/projected/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-kube-api-access-dwhst\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.889565 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.889555 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8b76471-1845-41af-aecd-ea80bf5ba97d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b8b76471-1845-41af-aecd-ea80bf5ba97d" (UID: "b8b76471-1845-41af-aecd-ea80bf5ba97d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.889788 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90c4575e-5eb6-4c24-8662-9c2c4863bc08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "90c4575e-5eb6-4c24-8662-9c2c4863bc08" (UID: "90c4575e-5eb6-4c24-8662-9c2c4863bc08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.892517 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8b76471-1845-41af-aecd-ea80bf5ba97d-kube-api-access-pm872" (OuterVolumeSpecName: "kube-api-access-pm872") pod "b8b76471-1845-41af-aecd-ea80bf5ba97d" (UID: "b8b76471-1845-41af-aecd-ea80bf5ba97d"). InnerVolumeSpecName "kube-api-access-pm872". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.892601 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90c4575e-5eb6-4c24-8662-9c2c4863bc08-kube-api-access-bbvt5" (OuterVolumeSpecName: "kube-api-access-bbvt5") pod "90c4575e-5eb6-4c24-8662-9c2c4863bc08" (UID: "90c4575e-5eb6-4c24-8662-9c2c4863bc08"). InnerVolumeSpecName "kube-api-access-bbvt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.991084 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm872\" (UniqueName: \"kubernetes.io/projected/b8b76471-1845-41af-aecd-ea80bf5ba97d-kube-api-access-pm872\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.991121 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b76471-1845-41af-aecd-ea80bf5ba97d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.991134 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbvt5\" (UniqueName: \"kubernetes.io/projected/90c4575e-5eb6-4c24-8662-9c2c4863bc08-kube-api-access-bbvt5\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.991146 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90c4575e-5eb6-4c24-8662-9c2c4863bc08-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.996673 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96e4-account-create-update-6fzmf" event={"ID":"90c4575e-5eb6-4c24-8662-9c2c4863bc08","Type":"ContainerDied","Data":"389f8010673b80b187f188cf0b94fb583f9b8377ca54dafb3e77df2f6f8e34d8"} Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.996714 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="389f8010673b80b187f188cf0b94fb583f9b8377ca54dafb3e77df2f6f8e34d8" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.996689 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96e4-account-create-update-6fzmf" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.998841 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-w4xpx" Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.998838 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-w4xpx" event={"ID":"f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee","Type":"ContainerDied","Data":"96e5d564ec6ac177b9ff81191ae02270d31e7c94affe78f818c3d10131e67b41"} Dec 05 08:34:54 crc kubenswrapper[4876]: I1205 08:34:54.999005 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96e5d564ec6ac177b9ff81191ae02270d31e7c94affe78f818c3d10131e67b41" Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.000698 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-348b-account-create-update-5dmk7" event={"ID":"2ee916a3-9c6b-4102-84b1-f921c334f152","Type":"ContainerDied","Data":"caa5713e485468b46bc0d0c39ad0747d52321484011832333e7b6ba42feedbec"} Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.000947 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-348b-account-create-update-5dmk7" Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.000954 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caa5713e485468b46bc0d0c39ad0747d52321484011832333e7b6ba42feedbec" Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.002330 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bwggg" Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.002325 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bwggg" event={"ID":"b8b76471-1845-41af-aecd-ea80bf5ba97d","Type":"ContainerDied","Data":"e01fa549d1ae1b66baf5762c1d6b06d7fcfab3035ec4b84ed994c5cec533779f"} Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.002448 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e01fa549d1ae1b66baf5762c1d6b06d7fcfab3035ec4b84ed994c5cec533779f" Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.002341 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-gs4d5" Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.066834 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gs4d5"] Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.076802 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gs4d5"] Dec 05 08:34:55 crc kubenswrapper[4876]: I1205 08:34:55.834208 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b95c35f4-164c-4466-954e-6005407caca5" path="/var/lib/kubelet/pods/b95c35f4-164c-4466-954e-6005407caca5/volumes" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.011707 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-mq6ht"] Dec 05 08:34:56 crc kubenswrapper[4876]: E1205 08:34:56.012784 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90c4575e-5eb6-4c24-8662-9c2c4863bc08" containerName="mariadb-account-create-update" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.012854 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="90c4575e-5eb6-4c24-8662-9c2c4863bc08" containerName="mariadb-account-create-update" Dec 05 08:34:56 crc kubenswrapper[4876]: E1205 08:34:56.012928 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee" containerName="mariadb-database-create" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.013004 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee" containerName="mariadb-database-create" Dec 05 08:34:56 crc kubenswrapper[4876]: E1205 08:34:56.013067 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8b76471-1845-41af-aecd-ea80bf5ba97d" containerName="mariadb-database-create" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.013124 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8b76471-1845-41af-aecd-ea80bf5ba97d" containerName="mariadb-database-create" Dec 05 08:34:56 crc kubenswrapper[4876]: E1205 08:34:56.013178 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee916a3-9c6b-4102-84b1-f921c334f152" containerName="mariadb-account-create-update" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.013231 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee916a3-9c6b-4102-84b1-f921c334f152" containerName="mariadb-account-create-update" Dec 05 08:34:56 crc kubenswrapper[4876]: E1205 08:34:56.013299 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95c35f4-164c-4466-954e-6005407caca5" containerName="dnsmasq-dns" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.013508 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95c35f4-164c-4466-954e-6005407caca5" containerName="dnsmasq-dns" Dec 05 08:34:56 crc kubenswrapper[4876]: E1205 08:34:56.013562 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b95c35f4-164c-4466-954e-6005407caca5" containerName="init" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.013617 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b95c35f4-164c-4466-954e-6005407caca5" containerName="init" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.013879 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="90c4575e-5eb6-4c24-8662-9c2c4863bc08" containerName="mariadb-account-create-update" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.013964 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee" containerName="mariadb-database-create" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.014023 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee916a3-9c6b-4102-84b1-f921c334f152" containerName="mariadb-account-create-update" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.014088 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8b76471-1845-41af-aecd-ea80bf5ba97d" containerName="mariadb-database-create" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.014147 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b95c35f4-164c-4466-954e-6005407caca5" containerName="dnsmasq-dns" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.014690 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.020329 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-mq6ht"] Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.138770 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-7dd9-account-create-update-k5wq2"] Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.139748 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.141755 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.152503 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7dd9-account-create-update-k5wq2"] Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.208526 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d158569-ccc7-4670-836b-c3459be0f527-operator-scripts\") pod \"glance-db-create-mq6ht\" (UID: \"4d158569-ccc7-4670-836b-c3459be0f527\") " pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.208711 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz565\" (UniqueName: \"kubernetes.io/projected/4d158569-ccc7-4670-836b-c3459be0f527-kube-api-access-sz565\") pod \"glance-db-create-mq6ht\" (UID: \"4d158569-ccc7-4670-836b-c3459be0f527\") " pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.310013 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d158569-ccc7-4670-836b-c3459be0f527-operator-scripts\") pod \"glance-db-create-mq6ht\" (UID: \"4d158569-ccc7-4670-836b-c3459be0f527\") " pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.310490 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz565\" (UniqueName: \"kubernetes.io/projected/4d158569-ccc7-4670-836b-c3459be0f527-kube-api-access-sz565\") pod \"glance-db-create-mq6ht\" (UID: \"4d158569-ccc7-4670-836b-c3459be0f527\") " pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.310814 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de5380be-9b7d-4320-ae64-a98efdafcb82-operator-scripts\") pod \"glance-7dd9-account-create-update-k5wq2\" (UID: \"de5380be-9b7d-4320-ae64-a98efdafcb82\") " pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.310970 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrsxz\" (UniqueName: \"kubernetes.io/projected/de5380be-9b7d-4320-ae64-a98efdafcb82-kube-api-access-nrsxz\") pod \"glance-7dd9-account-create-update-k5wq2\" (UID: \"de5380be-9b7d-4320-ae64-a98efdafcb82\") " pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.310719 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d158569-ccc7-4670-836b-c3459be0f527-operator-scripts\") pod \"glance-db-create-mq6ht\" (UID: \"4d158569-ccc7-4670-836b-c3459be0f527\") " pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.329356 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz565\" (UniqueName: \"kubernetes.io/projected/4d158569-ccc7-4670-836b-c3459be0f527-kube-api-access-sz565\") pod \"glance-db-create-mq6ht\" (UID: \"4d158569-ccc7-4670-836b-c3459be0f527\") " pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.413426 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de5380be-9b7d-4320-ae64-a98efdafcb82-operator-scripts\") pod \"glance-7dd9-account-create-update-k5wq2\" (UID: \"de5380be-9b7d-4320-ae64-a98efdafcb82\") " pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.414115 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrsxz\" (UniqueName: \"kubernetes.io/projected/de5380be-9b7d-4320-ae64-a98efdafcb82-kube-api-access-nrsxz\") pod \"glance-7dd9-account-create-update-k5wq2\" (UID: \"de5380be-9b7d-4320-ae64-a98efdafcb82\") " pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.414254 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de5380be-9b7d-4320-ae64-a98efdafcb82-operator-scripts\") pod \"glance-7dd9-account-create-update-k5wq2\" (UID: \"de5380be-9b7d-4320-ae64-a98efdafcb82\") " pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.431582 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrsxz\" (UniqueName: \"kubernetes.io/projected/de5380be-9b7d-4320-ae64-a98efdafcb82-kube-api-access-nrsxz\") pod \"glance-7dd9-account-create-update-k5wq2\" (UID: \"de5380be-9b7d-4320-ae64-a98efdafcb82\") " pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.455782 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.628958 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:56 crc kubenswrapper[4876]: I1205 08:34:56.697441 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7dd9-account-create-update-k5wq2"] Dec 05 08:34:56 crc kubenswrapper[4876]: W1205 08:34:56.702352 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde5380be_9b7d_4320_ae64_a98efdafcb82.slice/crio-e3bee043077a2d2e09b726ae68709363227b02f6c45f842b44360dacd09c2a3c WatchSource:0}: Error finding container e3bee043077a2d2e09b726ae68709363227b02f6c45f842b44360dacd09c2a3c: Status 404 returned error can't find the container with id e3bee043077a2d2e09b726ae68709363227b02f6c45f842b44360dacd09c2a3c Dec 05 08:34:57 crc kubenswrapper[4876]: I1205 08:34:57.021821 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7dd9-account-create-update-k5wq2" event={"ID":"de5380be-9b7d-4320-ae64-a98efdafcb82","Type":"ContainerStarted","Data":"e3bee043077a2d2e09b726ae68709363227b02f6c45f842b44360dacd09c2a3c"} Dec 05 08:34:57 crc kubenswrapper[4876]: I1205 08:34:57.081997 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-mq6ht"] Dec 05 08:34:57 crc kubenswrapper[4876]: W1205 08:34:57.085099 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d158569_ccc7_4670_836b_c3459be0f527.slice/crio-eda0c7191318b6a34be3ed01612342633269da976124dd3d5920466262a16216 WatchSource:0}: Error finding container eda0c7191318b6a34be3ed01612342633269da976124dd3d5920466262a16216: Status 404 returned error can't find the container with id eda0c7191318b6a34be3ed01612342633269da976124dd3d5920466262a16216 Dec 05 08:34:58 crc kubenswrapper[4876]: I1205 08:34:58.033991 4876 generic.go:334] "Generic (PLEG): container finished" podID="4d158569-ccc7-4670-836b-c3459be0f527" containerID="7b32f9d9dc611194a0a9ab986b1d4f84d0211cff6ec57beee41b83b486299032" exitCode=0 Dec 05 08:34:58 crc kubenswrapper[4876]: I1205 08:34:58.034047 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mq6ht" event={"ID":"4d158569-ccc7-4670-836b-c3459be0f527","Type":"ContainerDied","Data":"7b32f9d9dc611194a0a9ab986b1d4f84d0211cff6ec57beee41b83b486299032"} Dec 05 08:34:58 crc kubenswrapper[4876]: I1205 08:34:58.034771 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mq6ht" event={"ID":"4d158569-ccc7-4670-836b-c3459be0f527","Type":"ContainerStarted","Data":"eda0c7191318b6a34be3ed01612342633269da976124dd3d5920466262a16216"} Dec 05 08:34:58 crc kubenswrapper[4876]: I1205 08:34:58.036775 4876 generic.go:334] "Generic (PLEG): container finished" podID="de5380be-9b7d-4320-ae64-a98efdafcb82" containerID="3eb6f5ab627df8b351bd9b5d1b6b6df495fc5d8862ae6dee1286d23f97eebf8f" exitCode=0 Dec 05 08:34:58 crc kubenswrapper[4876]: I1205 08:34:58.036823 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7dd9-account-create-update-k5wq2" event={"ID":"de5380be-9b7d-4320-ae64-a98efdafcb82","Type":"ContainerDied","Data":"3eb6f5ab627df8b351bd9b5d1b6b6df495fc5d8862ae6dee1286d23f97eebf8f"} Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.052321 4876 generic.go:334] "Generic (PLEG): container finished" podID="d41c7aad-b564-4c13-9f60-2d1fe53861a8" containerID="25f444d0078306eb58644ca66e8c564a9c14c9c8d23d12c8acc2f46026721a6f" exitCode=0 Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.052454 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hzjsf" event={"ID":"d41c7aad-b564-4c13-9f60-2d1fe53861a8","Type":"ContainerDied","Data":"25f444d0078306eb58644ca66e8c564a9c14c9c8d23d12c8acc2f46026721a6f"} Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.468156 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mq6ht" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.479856 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.515036 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d158569-ccc7-4670-836b-c3459be0f527-operator-scripts\") pod \"4d158569-ccc7-4670-836b-c3459be0f527\" (UID: \"4d158569-ccc7-4670-836b-c3459be0f527\") " Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.515137 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz565\" (UniqueName: \"kubernetes.io/projected/4d158569-ccc7-4670-836b-c3459be0f527-kube-api-access-sz565\") pod \"4d158569-ccc7-4670-836b-c3459be0f527\" (UID: \"4d158569-ccc7-4670-836b-c3459be0f527\") " Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.515948 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d158569-ccc7-4670-836b-c3459be0f527-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d158569-ccc7-4670-836b-c3459be0f527" (UID: "4d158569-ccc7-4670-836b-c3459be0f527"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.521771 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d158569-ccc7-4670-836b-c3459be0f527-kube-api-access-sz565" (OuterVolumeSpecName: "kube-api-access-sz565") pod "4d158569-ccc7-4670-836b-c3459be0f527" (UID: "4d158569-ccc7-4670-836b-c3459be0f527"). InnerVolumeSpecName "kube-api-access-sz565". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.616411 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de5380be-9b7d-4320-ae64-a98efdafcb82-operator-scripts\") pod \"de5380be-9b7d-4320-ae64-a98efdafcb82\" (UID: \"de5380be-9b7d-4320-ae64-a98efdafcb82\") " Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.616581 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrsxz\" (UniqueName: \"kubernetes.io/projected/de5380be-9b7d-4320-ae64-a98efdafcb82-kube-api-access-nrsxz\") pod \"de5380be-9b7d-4320-ae64-a98efdafcb82\" (UID: \"de5380be-9b7d-4320-ae64-a98efdafcb82\") " Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.617387 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d158569-ccc7-4670-836b-c3459be0f527-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.617410 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz565\" (UniqueName: \"kubernetes.io/projected/4d158569-ccc7-4670-836b-c3459be0f527-kube-api-access-sz565\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.618050 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de5380be-9b7d-4320-ae64-a98efdafcb82-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "de5380be-9b7d-4320-ae64-a98efdafcb82" (UID: "de5380be-9b7d-4320-ae64-a98efdafcb82"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.622460 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de5380be-9b7d-4320-ae64-a98efdafcb82-kube-api-access-nrsxz" (OuterVolumeSpecName: "kube-api-access-nrsxz") pod "de5380be-9b7d-4320-ae64-a98efdafcb82" (UID: "de5380be-9b7d-4320-ae64-a98efdafcb82"). InnerVolumeSpecName "kube-api-access-nrsxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.720998 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.721186 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de5380be-9b7d-4320-ae64-a98efdafcb82-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.721206 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrsxz\" (UniqueName: \"kubernetes.io/projected/de5380be-9b7d-4320-ae64-a98efdafcb82-kube-api-access-nrsxz\") on node \"crc\" DevicePath \"\"" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.727956 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71735818-3684-481f-bbb3-72aff2d50428-etc-swift\") pod \"swift-storage-0\" (UID: \"71735818-3684-481f-bbb3-72aff2d50428\") " pod="openstack/swift-storage-0" Dec 05 08:34:59 crc kubenswrapper[4876]: I1205 08:34:59.977578 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.071415 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7dd9-account-create-update-k5wq2" event={"ID":"de5380be-9b7d-4320-ae64-a98efdafcb82","Type":"ContainerDied","Data":"e3bee043077a2d2e09b726ae68709363227b02f6c45f842b44360dacd09c2a3c"} Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.071455 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3bee043077a2d2e09b726ae68709363227b02f6c45f842b44360dacd09c2a3c" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.071474 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7dd9-account-create-update-k5wq2" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.091277 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-mq6ht" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.091329 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-mq6ht" event={"ID":"4d158569-ccc7-4670-836b-c3459be0f527","Type":"ContainerDied","Data":"eda0c7191318b6a34be3ed01612342633269da976124dd3d5920466262a16216"} Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.091376 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eda0c7191318b6a34be3ed01612342633269da976124dd3d5920466262a16216" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.560179 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:35:00 crc kubenswrapper[4876]: W1205 08:35:00.598581 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71735818_3684_481f_bbb3_72aff2d50428.slice/crio-bbb6df06f490d6d1b32f37dde662e032d9f0c9d2cbfc8162e078e6849c93e663 WatchSource:0}: Error finding container bbb6df06f490d6d1b32f37dde662e032d9f0c9d2cbfc8162e078e6849c93e663: Status 404 returned error can't find the container with id bbb6df06f490d6d1b32f37dde662e032d9f0c9d2cbfc8162e078e6849c93e663 Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.606762 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.642613 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzgb8\" (UniqueName: \"kubernetes.io/projected/d41c7aad-b564-4c13-9f60-2d1fe53861a8-kube-api-access-gzgb8\") pod \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.642664 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-combined-ca-bundle\") pod \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.642740 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d41c7aad-b564-4c13-9f60-2d1fe53861a8-etc-swift\") pod \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.642768 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-swiftconf\") pod \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.642845 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-ring-data-devices\") pod \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.642889 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-dispersionconf\") pod \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.642947 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-scripts\") pod \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\" (UID: \"d41c7aad-b564-4c13-9f60-2d1fe53861a8\") " Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.643690 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d41c7aad-b564-4c13-9f60-2d1fe53861a8" (UID: "d41c7aad-b564-4c13-9f60-2d1fe53861a8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.644045 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d41c7aad-b564-4c13-9f60-2d1fe53861a8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d41c7aad-b564-4c13-9f60-2d1fe53861a8" (UID: "d41c7aad-b564-4c13-9f60-2d1fe53861a8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.648261 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d41c7aad-b564-4c13-9f60-2d1fe53861a8-kube-api-access-gzgb8" (OuterVolumeSpecName: "kube-api-access-gzgb8") pod "d41c7aad-b564-4c13-9f60-2d1fe53861a8" (UID: "d41c7aad-b564-4c13-9f60-2d1fe53861a8"). InnerVolumeSpecName "kube-api-access-gzgb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.655295 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d41c7aad-b564-4c13-9f60-2d1fe53861a8" (UID: "d41c7aad-b564-4c13-9f60-2d1fe53861a8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.667956 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d41c7aad-b564-4c13-9f60-2d1fe53861a8" (UID: "d41c7aad-b564-4c13-9f60-2d1fe53861a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.671458 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-scripts" (OuterVolumeSpecName: "scripts") pod "d41c7aad-b564-4c13-9f60-2d1fe53861a8" (UID: "d41c7aad-b564-4c13-9f60-2d1fe53861a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.671483 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d41c7aad-b564-4c13-9f60-2d1fe53861a8" (UID: "d41c7aad-b564-4c13-9f60-2d1fe53861a8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.744527 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.744570 4876 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d41c7aad-b564-4c13-9f60-2d1fe53861a8-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.744584 4876 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.744595 4876 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.744607 4876 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d41c7aad-b564-4c13-9f60-2d1fe53861a8-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.744618 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d41c7aad-b564-4c13-9f60-2d1fe53861a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.744629 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzgb8\" (UniqueName: \"kubernetes.io/projected/d41c7aad-b564-4c13-9f60-2d1fe53861a8-kube-api-access-gzgb8\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.805548 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-7p5wk" podUID="442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2" containerName="ovn-controller" probeResult="failure" output=< Dec 05 08:35:00 crc kubenswrapper[4876]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 08:35:00 crc kubenswrapper[4876]: > Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.879085 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:35:00 crc kubenswrapper[4876]: I1205 08:35:00.884274 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-fhwc7" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.100233 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7p5wk-config-gxjt8"] Dec 05 08:35:01 crc kubenswrapper[4876]: E1205 08:35:01.100549 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d158569-ccc7-4670-836b-c3459be0f527" containerName="mariadb-database-create" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.100561 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d158569-ccc7-4670-836b-c3459be0f527" containerName="mariadb-database-create" Dec 05 08:35:01 crc kubenswrapper[4876]: E1205 08:35:01.100575 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de5380be-9b7d-4320-ae64-a98efdafcb82" containerName="mariadb-account-create-update" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.100581 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="de5380be-9b7d-4320-ae64-a98efdafcb82" containerName="mariadb-account-create-update" Dec 05 08:35:01 crc kubenswrapper[4876]: E1205 08:35:01.100594 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d41c7aad-b564-4c13-9f60-2d1fe53861a8" containerName="swift-ring-rebalance" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.100601 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d41c7aad-b564-4c13-9f60-2d1fe53861a8" containerName="swift-ring-rebalance" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.100770 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d158569-ccc7-4670-836b-c3459be0f527" containerName="mariadb-database-create" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.100787 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d41c7aad-b564-4c13-9f60-2d1fe53861a8" containerName="swift-ring-rebalance" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.100795 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="de5380be-9b7d-4320-ae64-a98efdafcb82" containerName="mariadb-account-create-update" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.101289 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.103629 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.106003 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hzjsf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.105995 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hzjsf" event={"ID":"d41c7aad-b564-4c13-9f60-2d1fe53861a8","Type":"ContainerDied","Data":"a816b20455147ccd217afa20c75379d002dd7119516aeb464157a934d5648549"} Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.106227 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a816b20455147ccd217afa20c75379d002dd7119516aeb464157a934d5648549" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.117994 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"bbb6df06f490d6d1b32f37dde662e032d9f0c9d2cbfc8162e078e6849c93e663"} Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.125769 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7p5wk-config-gxjt8"] Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.151096 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-additional-scripts\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.151375 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.151571 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-log-ovn\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.151812 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsn4m\" (UniqueName: \"kubernetes.io/projected/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-kube-api-access-zsn4m\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.152023 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run-ovn\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.152236 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-scripts\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.253397 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-log-ovn\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.253490 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsn4m\" (UniqueName: \"kubernetes.io/projected/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-kube-api-access-zsn4m\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.253561 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run-ovn\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.253584 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-scripts\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.253611 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-additional-scripts\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.253726 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.254348 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.254412 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-log-ovn\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.254984 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run-ovn\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.255917 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-additional-scripts\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.257018 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-scripts\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.277843 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsn4m\" (UniqueName: \"kubernetes.io/projected/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-kube-api-access-zsn4m\") pod \"ovn-controller-7p5wk-config-gxjt8\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.420314 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-n7phf"] Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.421799 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.426302 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5msdq" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.426697 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.428814 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.432640 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-n7phf"] Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.558854 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-db-sync-config-data\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.558923 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-config-data\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.559061 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d92bv\" (UniqueName: \"kubernetes.io/projected/2c58acae-5113-4e33-951e-409855923e87-kube-api-access-d92bv\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.559113 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-combined-ca-bundle\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.663796 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d92bv\" (UniqueName: \"kubernetes.io/projected/2c58acae-5113-4e33-951e-409855923e87-kube-api-access-d92bv\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.663894 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-combined-ca-bundle\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.663963 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-db-sync-config-data\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.663984 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-config-data\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.669509 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-config-data\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.669737 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-combined-ca-bundle\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.685449 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-db-sync-config-data\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.690134 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d92bv\" (UniqueName: \"kubernetes.io/projected/2c58acae-5113-4e33-951e-409855923e87-kube-api-access-d92bv\") pod \"glance-db-sync-n7phf\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:01 crc kubenswrapper[4876]: I1205 08:35:01.825053 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:02 crc kubenswrapper[4876]: I1205 08:35:02.004263 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7p5wk-config-gxjt8"] Dec 05 08:35:02 crc kubenswrapper[4876]: I1205 08:35:02.138679 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7p5wk-config-gxjt8" event={"ID":"cbd29938-b5a1-4e36-b585-15bd27eb4c6d","Type":"ContainerStarted","Data":"54dd8dcb2cbf19fcb097ed5709e03b49156de109ab111fd6f709277cd760e67a"} Dec 05 08:35:02 crc kubenswrapper[4876]: I1205 08:35:02.370980 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-n7phf"] Dec 05 08:35:02 crc kubenswrapper[4876]: W1205 08:35:02.373738 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c58acae_5113_4e33_951e_409855923e87.slice/crio-19da0bd05f039a884a5201c7d5280ddf3345d22fa708081943e11272e74377e0 WatchSource:0}: Error finding container 19da0bd05f039a884a5201c7d5280ddf3345d22fa708081943e11272e74377e0: Status 404 returned error can't find the container with id 19da0bd05f039a884a5201c7d5280ddf3345d22fa708081943e11272e74377e0 Dec 05 08:35:03 crc kubenswrapper[4876]: I1205 08:35:03.147989 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n7phf" event={"ID":"2c58acae-5113-4e33-951e-409855923e87","Type":"ContainerStarted","Data":"19da0bd05f039a884a5201c7d5280ddf3345d22fa708081943e11272e74377e0"} Dec 05 08:35:03 crc kubenswrapper[4876]: I1205 08:35:03.152463 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7p5wk-config-gxjt8" event={"ID":"cbd29938-b5a1-4e36-b585-15bd27eb4c6d","Type":"ContainerStarted","Data":"baae878ad4ed798eea62443cd4ff0f9e29cbf336ba66ef9751d47a6ae8ade703"} Dec 05 08:35:03 crc kubenswrapper[4876]: I1205 08:35:03.173384 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-7p5wk-config-gxjt8" podStartSLOduration=2.173364773 podStartE2EDuration="2.173364773s" podCreationTimestamp="2025-12-05 08:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:03.167854637 +0000 UTC m=+1167.656519299" watchObservedRunningTime="2025-12-05 08:35:03.173364773 +0000 UTC m=+1167.662029395" Dec 05 08:35:04 crc kubenswrapper[4876]: I1205 08:35:04.165717 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"1012dc487f337a400c93c3d5cbf7cf0da8ec4917dcae7380c804f865007761b4"} Dec 05 08:35:04 crc kubenswrapper[4876]: I1205 08:35:04.166022 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"55edf6bc092c09de1b513d2a3e6467bc6c9412c586ae0cbfbaf01836f8188b02"} Dec 05 08:35:04 crc kubenswrapper[4876]: I1205 08:35:04.166033 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"ba1b867c98f02f8894953b4f80ac4099b5d368c87e9fc35579d19ed798db4e46"} Dec 05 08:35:04 crc kubenswrapper[4876]: I1205 08:35:04.172595 4876 generic.go:334] "Generic (PLEG): container finished" podID="cbd29938-b5a1-4e36-b585-15bd27eb4c6d" containerID="baae878ad4ed798eea62443cd4ff0f9e29cbf336ba66ef9751d47a6ae8ade703" exitCode=0 Dec 05 08:35:04 crc kubenswrapper[4876]: I1205 08:35:04.172700 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7p5wk-config-gxjt8" event={"ID":"cbd29938-b5a1-4e36-b585-15bd27eb4c6d","Type":"ContainerDied","Data":"baae878ad4ed798eea62443cd4ff0f9e29cbf336ba66ef9751d47a6ae8ade703"} Dec 05 08:35:04 crc kubenswrapper[4876]: I1205 08:35:04.174175 4876 generic.go:334] "Generic (PLEG): container finished" podID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerID="f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369" exitCode=0 Dec 05 08:35:04 crc kubenswrapper[4876]: I1205 08:35:04.174204 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45b18655-19ec-43f4-a779-b8f72bc77aa2","Type":"ContainerDied","Data":"f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369"} Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.186827 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"f542ce36a6bceb28390b2fe1decebe9a0ecdbad3f2783f1ca40fa8849aba485d"} Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.189524 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45b18655-19ec-43f4-a779-b8f72bc77aa2","Type":"ContainerStarted","Data":"a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6"} Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.189762 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.227435 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.420460054 podStartE2EDuration="1m19.227417474s" podCreationTimestamp="2025-12-05 08:33:46 +0000 UTC" firstStartedPulling="2025-12-05 08:33:49.148993676 +0000 UTC m=+1093.637658298" lastFinishedPulling="2025-12-05 08:34:27.955951096 +0000 UTC m=+1132.444615718" observedRunningTime="2025-12-05 08:35:05.227018042 +0000 UTC m=+1169.715682664" watchObservedRunningTime="2025-12-05 08:35:05.227417474 +0000 UTC m=+1169.716082096" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.635580 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683081 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run-ovn\") pod \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683197 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-log-ovn\") pod \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683231 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run\") pod \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683260 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsn4m\" (UniqueName: \"kubernetes.io/projected/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-kube-api-access-zsn4m\") pod \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683343 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-scripts\") pod \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683456 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-additional-scripts\") pod \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\" (UID: \"cbd29938-b5a1-4e36-b585-15bd27eb4c6d\") " Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683513 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "cbd29938-b5a1-4e36-b585-15bd27eb4c6d" (UID: "cbd29938-b5a1-4e36-b585-15bd27eb4c6d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683593 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run" (OuterVolumeSpecName: "var-run") pod "cbd29938-b5a1-4e36-b585-15bd27eb4c6d" (UID: "cbd29938-b5a1-4e36-b585-15bd27eb4c6d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.683647 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "cbd29938-b5a1-4e36-b585-15bd27eb4c6d" (UID: "cbd29938-b5a1-4e36-b585-15bd27eb4c6d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.684334 4876 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.684406 4876 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.684491 4876 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.684719 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "cbd29938-b5a1-4e36-b585-15bd27eb4c6d" (UID: "cbd29938-b5a1-4e36-b585-15bd27eb4c6d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.684773 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-scripts" (OuterVolumeSpecName: "scripts") pod "cbd29938-b5a1-4e36-b585-15bd27eb4c6d" (UID: "cbd29938-b5a1-4e36-b585-15bd27eb4c6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.692049 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-kube-api-access-zsn4m" (OuterVolumeSpecName: "kube-api-access-zsn4m") pod "cbd29938-b5a1-4e36-b585-15bd27eb4c6d" (UID: "cbd29938-b5a1-4e36-b585-15bd27eb4c6d"). InnerVolumeSpecName "kube-api-access-zsn4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.787497 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsn4m\" (UniqueName: \"kubernetes.io/projected/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-kube-api-access-zsn4m\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.787536 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.787546 4876 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd29938-b5a1-4e36-b585-15bd27eb4c6d-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:05 crc kubenswrapper[4876]: I1205 08:35:05.812117 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-7p5wk" Dec 05 08:35:06 crc kubenswrapper[4876]: I1205 08:35:06.208273 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7p5wk-config-gxjt8" Dec 05 08:35:06 crc kubenswrapper[4876]: I1205 08:35:06.208987 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7p5wk-config-gxjt8" event={"ID":"cbd29938-b5a1-4e36-b585-15bd27eb4c6d","Type":"ContainerDied","Data":"54dd8dcb2cbf19fcb097ed5709e03b49156de109ab111fd6f709277cd760e67a"} Dec 05 08:35:06 crc kubenswrapper[4876]: I1205 08:35:06.209008 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54dd8dcb2cbf19fcb097ed5709e03b49156de109ab111fd6f709277cd760e67a" Dec 05 08:35:06 crc kubenswrapper[4876]: I1205 08:35:06.290577 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-7p5wk-config-gxjt8"] Dec 05 08:35:06 crc kubenswrapper[4876]: I1205 08:35:06.298291 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-7p5wk-config-gxjt8"] Dec 05 08:35:07 crc kubenswrapper[4876]: I1205 08:35:07.225620 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"10fb4495deca3de7c1ef4588743d921afc48e6809f62b2cb5110ecf61cb54a2c"} Dec 05 08:35:07 crc kubenswrapper[4876]: I1205 08:35:07.225979 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"2751e0f4777874da2ebb0f364b53fa6e14dd5aa347abbafa8f4bdd24d058fd18"} Dec 05 08:35:07 crc kubenswrapper[4876]: I1205 08:35:07.838600 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbd29938-b5a1-4e36-b585-15bd27eb4c6d" path="/var/lib/kubelet/pods/cbd29938-b5a1-4e36-b585-15bd27eb4c6d/volumes" Dec 05 08:35:08 crc kubenswrapper[4876]: I1205 08:35:08.236988 4876 generic.go:334] "Generic (PLEG): container finished" podID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerID="59bcde78d239d2ec5b2c1b6c42bfe979ce595826cff0ba2f07bdfb82de0dac70" exitCode=0 Dec 05 08:35:08 crc kubenswrapper[4876]: I1205 08:35:08.237049 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f","Type":"ContainerDied","Data":"59bcde78d239d2ec5b2c1b6c42bfe979ce595826cff0ba2f07bdfb82de0dac70"} Dec 05 08:35:08 crc kubenswrapper[4876]: I1205 08:35:08.259768 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"f1b57a3e1e41b7d62c7d6846398f5417c7bebb393a031f44b23c17c72ec0ecc3"} Dec 05 08:35:08 crc kubenswrapper[4876]: I1205 08:35:08.259819 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"f5deded80ece8f944b9ebca7d36e7bad0fd6f486a2af1f441479d9a467a17698"} Dec 05 08:35:09 crc kubenswrapper[4876]: I1205 08:35:09.334754 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"dd563abccff3bf7d0802ac0c722cf446ae2b0a6bec8852fee8a6dc85ce6be2b4"} Dec 05 08:35:09 crc kubenswrapper[4876]: I1205 08:35:09.335305 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"ab938f31e71a433542115679cc5f84f994146989b8f95b527eff1fe05f46f933"} Dec 05 08:35:09 crc kubenswrapper[4876]: I1205 08:35:09.338631 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f","Type":"ContainerStarted","Data":"6cc496e7c01898528567beed9d06b8072750b24382edf8aea2e624e6163bbd2b"} Dec 05 08:35:09 crc kubenswrapper[4876]: I1205 08:35:09.338836 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 08:35:09 crc kubenswrapper[4876]: I1205 08:35:09.365136 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371953.489655 podStartE2EDuration="1m23.365121619s" podCreationTimestamp="2025-12-05 08:33:46 +0000 UTC" firstStartedPulling="2025-12-05 08:33:48.230426743 +0000 UTC m=+1092.719091365" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:09.361765554 +0000 UTC m=+1173.850430196" watchObservedRunningTime="2025-12-05 08:35:09.365121619 +0000 UTC m=+1173.853786241" Dec 05 08:35:10 crc kubenswrapper[4876]: I1205 08:35:10.357962 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"cb2ce6e7ecd7afa5692d0362b1b0cb9bc8204d8ddc074562edf2b19eaa45e2be"} Dec 05 08:35:10 crc kubenswrapper[4876]: I1205 08:35:10.358001 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"8bc629da66ddad6803e18d78e51b5e9649cad011adf68eb0c7bdf64b12df364d"} Dec 05 08:35:10 crc kubenswrapper[4876]: I1205 08:35:10.358010 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"3a3a187edeff616770e469d7c02a116fbf4ffef2b28322f5cfcf00d56d964065"} Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.372520 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"47622a1e7ae442ce4f2377429de62ad428508e8cf5f60c4751b788a84980a8ae"} Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.372992 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"71735818-3684-481f-bbb3-72aff2d50428","Type":"ContainerStarted","Data":"50a4c05e8e63de6ef2c7f2996102b945d8606a8959bf0696431e52fc606eabf5"} Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.438112 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.213538312 podStartE2EDuration="29.438097575s" podCreationTimestamp="2025-12-05 08:34:42 +0000 UTC" firstStartedPulling="2025-12-05 08:35:00.601376242 +0000 UTC m=+1165.090040884" lastFinishedPulling="2025-12-05 08:35:08.825935495 +0000 UTC m=+1173.314600147" observedRunningTime="2025-12-05 08:35:11.433026962 +0000 UTC m=+1175.921691594" watchObservedRunningTime="2025-12-05 08:35:11.438097575 +0000 UTC m=+1175.926762197" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.782963 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hchbt"] Dec 05 08:35:11 crc kubenswrapper[4876]: E1205 08:35:11.783302 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd29938-b5a1-4e36-b585-15bd27eb4c6d" containerName="ovn-config" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.783318 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd29938-b5a1-4e36-b585-15bd27eb4c6d" containerName="ovn-config" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.783490 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbd29938-b5a1-4e36-b585-15bd27eb4c6d" containerName="ovn-config" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.784333 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.787191 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.797853 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hchbt"] Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.811390 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.811519 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.811569 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.811606 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxf5f\" (UniqueName: \"kubernetes.io/projected/6e7c039f-5af6-4394-b41b-9d8273912ddb-kube-api-access-fxf5f\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.811666 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.811701 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-config\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.913282 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.913669 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxf5f\" (UniqueName: \"kubernetes.io/projected/6e7c039f-5af6-4394-b41b-9d8273912ddb-kube-api-access-fxf5f\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.913706 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.913730 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-config\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.913773 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.914587 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.914692 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.914803 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-config\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.914924 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.914942 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.915612 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:11 crc kubenswrapper[4876]: I1205 08:35:11.954063 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxf5f\" (UniqueName: \"kubernetes.io/projected/6e7c039f-5af6-4394-b41b-9d8273912ddb-kube-api-access-fxf5f\") pod \"dnsmasq-dns-6d5b6d6b67-hchbt\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:12 crc kubenswrapper[4876]: I1205 08:35:12.103376 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:18 crc kubenswrapper[4876]: I1205 08:35:18.092126 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:35:20 crc kubenswrapper[4876]: I1205 08:35:20.817311 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hchbt"] Dec 05 08:35:20 crc kubenswrapper[4876]: W1205 08:35:20.826223 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e7c039f_5af6_4394_b41b_9d8273912ddb.slice/crio-b1f0b3239dec7e77ec5389ad5563909a5ef6e206910bf7744efba04c0232a8bd WatchSource:0}: Error finding container b1f0b3239dec7e77ec5389ad5563909a5ef6e206910bf7744efba04c0232a8bd: Status 404 returned error can't find the container with id b1f0b3239dec7e77ec5389ad5563909a5ef6e206910bf7744efba04c0232a8bd Dec 05 08:35:21 crc kubenswrapper[4876]: I1205 08:35:21.630346 4876 generic.go:334] "Generic (PLEG): container finished" podID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerID="c6703baa6ac24ed7638811d50e540df6ad4972e57ef9a9723c2ff56664b78e38" exitCode=0 Dec 05 08:35:21 crc kubenswrapper[4876]: I1205 08:35:21.631156 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" event={"ID":"6e7c039f-5af6-4394-b41b-9d8273912ddb","Type":"ContainerDied","Data":"c6703baa6ac24ed7638811d50e540df6ad4972e57ef9a9723c2ff56664b78e38"} Dec 05 08:35:21 crc kubenswrapper[4876]: I1205 08:35:21.631229 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" event={"ID":"6e7c039f-5af6-4394-b41b-9d8273912ddb","Type":"ContainerStarted","Data":"b1f0b3239dec7e77ec5389ad5563909a5ef6e206910bf7744efba04c0232a8bd"} Dec 05 08:35:21 crc kubenswrapper[4876]: I1205 08:35:21.634183 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n7phf" event={"ID":"2c58acae-5113-4e33-951e-409855923e87","Type":"ContainerStarted","Data":"bfe4c276c30394787954befcccf806035dc8eef3b38dddf3dc8b4e446b5807c4"} Dec 05 08:35:21 crc kubenswrapper[4876]: I1205 08:35:21.696892 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-n7phf" podStartSLOduration=2.5968655800000002 podStartE2EDuration="20.696874451s" podCreationTimestamp="2025-12-05 08:35:01 +0000 UTC" firstStartedPulling="2025-12-05 08:35:02.376379373 +0000 UTC m=+1166.865043995" lastFinishedPulling="2025-12-05 08:35:20.476388244 +0000 UTC m=+1184.965052866" observedRunningTime="2025-12-05 08:35:21.688161105 +0000 UTC m=+1186.176825737" watchObservedRunningTime="2025-12-05 08:35:21.696874451 +0000 UTC m=+1186.185539073" Dec 05 08:35:22 crc kubenswrapper[4876]: I1205 08:35:22.645784 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" event={"ID":"6e7c039f-5af6-4394-b41b-9d8273912ddb","Type":"ContainerStarted","Data":"08e18a87af24b5174714da524d5711889d67739069c67b03d6397d683361086b"} Dec 05 08:35:22 crc kubenswrapper[4876]: I1205 08:35:22.646273 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:22 crc kubenswrapper[4876]: I1205 08:35:22.679181 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" podStartSLOduration=11.679154887 podStartE2EDuration="11.679154887s" podCreationTimestamp="2025-12-05 08:35:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:22.668840306 +0000 UTC m=+1187.157505018" watchObservedRunningTime="2025-12-05 08:35:22.679154887 +0000 UTC m=+1187.167819549" Dec 05 08:35:27 crc kubenswrapper[4876]: I1205 08:35:27.131198 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:35:27 crc kubenswrapper[4876]: I1205 08:35:27.198407 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gvpkj"] Dec 05 08:35:27 crc kubenswrapper[4876]: I1205 08:35:27.198930 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" podUID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerName="dnsmasq-dns" containerID="cri-o://3a63fd7d71efca11be3d44d274811be02f2758a0769977e828e60ba2761b9b43" gracePeriod=10 Dec 05 08:35:27 crc kubenswrapper[4876]: I1205 08:35:27.439174 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 08:35:27 crc kubenswrapper[4876]: I1205 08:35:27.870989 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-mm4br"] Dec 05 08:35:27 crc kubenswrapper[4876]: I1205 08:35:27.872703 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:27 crc kubenswrapper[4876]: I1205 08:35:27.884951 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mm4br"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.112085 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db2de57d-f90f-44c1-89de-d7706bd6e64c-operator-scripts\") pod \"barbican-db-create-mm4br\" (UID: \"db2de57d-f90f-44c1-89de-d7706bd6e64c\") " pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.112213 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwvgm\" (UniqueName: \"kubernetes.io/projected/db2de57d-f90f-44c1-89de-d7706bd6e64c-kube-api-access-nwvgm\") pod \"barbican-db-create-mm4br\" (UID: \"db2de57d-f90f-44c1-89de-d7706bd6e64c\") " pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.150062 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6e8e-account-create-update-jnkdv"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.151106 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.153101 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.163579 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6e8e-account-create-update-jnkdv"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.213275 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db2de57d-f90f-44c1-89de-d7706bd6e64c-operator-scripts\") pod \"barbican-db-create-mm4br\" (UID: \"db2de57d-f90f-44c1-89de-d7706bd6e64c\") " pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.213516 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwvgm\" (UniqueName: \"kubernetes.io/projected/db2de57d-f90f-44c1-89de-d7706bd6e64c-kube-api-access-nwvgm\") pod \"barbican-db-create-mm4br\" (UID: \"db2de57d-f90f-44c1-89de-d7706bd6e64c\") " pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.213647 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-676pp\" (UniqueName: \"kubernetes.io/projected/9855cf16-5f53-4795-9103-3ee2c6b75f79-kube-api-access-676pp\") pod \"cinder-6e8e-account-create-update-jnkdv\" (UID: \"9855cf16-5f53-4795-9103-3ee2c6b75f79\") " pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.213781 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9855cf16-5f53-4795-9103-3ee2c6b75f79-operator-scripts\") pod \"cinder-6e8e-account-create-update-jnkdv\" (UID: \"9855cf16-5f53-4795-9103-3ee2c6b75f79\") " pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.214182 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db2de57d-f90f-44c1-89de-d7706bd6e64c-operator-scripts\") pod \"barbican-db-create-mm4br\" (UID: \"db2de57d-f90f-44c1-89de-d7706bd6e64c\") " pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.246923 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-bzms5"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.248269 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bzms5" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.251417 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" podUID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.251678 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwvgm\" (UniqueName: \"kubernetes.io/projected/db2de57d-f90f-44c1-89de-d7706bd6e64c-kube-api-access-nwvgm\") pod \"barbican-db-create-mm4br\" (UID: \"db2de57d-f90f-44c1-89de-d7706bd6e64c\") " pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.269075 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5909-account-create-update-6pw92"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.270146 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.271936 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.279008 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-bzms5"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.304611 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5909-account-create-update-6pw92"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.315108 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkv2w\" (UniqueName: \"kubernetes.io/projected/80e46c2c-ff07-4431-ab74-fafdc1581673-kube-api-access-nkv2w\") pod \"barbican-5909-account-create-update-6pw92\" (UID: \"80e46c2c-ff07-4431-ab74-fafdc1581673\") " pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.315155 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n5kp\" (UniqueName: \"kubernetes.io/projected/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-kube-api-access-5n5kp\") pod \"heat-db-create-bzms5\" (UID: \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\") " pod="openstack/heat-db-create-bzms5" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.315186 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e46c2c-ff07-4431-ab74-fafdc1581673-operator-scripts\") pod \"barbican-5909-account-create-update-6pw92\" (UID: \"80e46c2c-ff07-4431-ab74-fafdc1581673\") " pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.315212 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-676pp\" (UniqueName: \"kubernetes.io/projected/9855cf16-5f53-4795-9103-3ee2c6b75f79-kube-api-access-676pp\") pod \"cinder-6e8e-account-create-update-jnkdv\" (UID: \"9855cf16-5f53-4795-9103-3ee2c6b75f79\") " pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.315247 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9855cf16-5f53-4795-9103-3ee2c6b75f79-operator-scripts\") pod \"cinder-6e8e-account-create-update-jnkdv\" (UID: \"9855cf16-5f53-4795-9103-3ee2c6b75f79\") " pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.315285 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-operator-scripts\") pod \"heat-db-create-bzms5\" (UID: \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\") " pod="openstack/heat-db-create-bzms5" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.316336 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9855cf16-5f53-4795-9103-3ee2c6b75f79-operator-scripts\") pod \"cinder-6e8e-account-create-update-jnkdv\" (UID: \"9855cf16-5f53-4795-9103-3ee2c6b75f79\") " pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.345077 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-676pp\" (UniqueName: \"kubernetes.io/projected/9855cf16-5f53-4795-9103-3ee2c6b75f79-kube-api-access-676pp\") pod \"cinder-6e8e-account-create-update-jnkdv\" (UID: \"9855cf16-5f53-4795-9103-3ee2c6b75f79\") " pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.353800 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-xz6wd"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.355824 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.367286 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xz6wd"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.421326 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n5kp\" (UniqueName: \"kubernetes.io/projected/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-kube-api-access-5n5kp\") pod \"heat-db-create-bzms5\" (UID: \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\") " pod="openstack/heat-db-create-bzms5" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.421767 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e46c2c-ff07-4431-ab74-fafdc1581673-operator-scripts\") pod \"barbican-5909-account-create-update-6pw92\" (UID: \"80e46c2c-ff07-4431-ab74-fafdc1581673\") " pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.421877 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-operator-scripts\") pod \"heat-db-create-bzms5\" (UID: \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\") " pod="openstack/heat-db-create-bzms5" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.422025 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkv2w\" (UniqueName: \"kubernetes.io/projected/80e46c2c-ff07-4431-ab74-fafdc1581673-kube-api-access-nkv2w\") pod \"barbican-5909-account-create-update-6pw92\" (UID: \"80e46c2c-ff07-4431-ab74-fafdc1581673\") " pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.423116 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e46c2c-ff07-4431-ab74-fafdc1581673-operator-scripts\") pod \"barbican-5909-account-create-update-6pw92\" (UID: \"80e46c2c-ff07-4431-ab74-fafdc1581673\") " pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.423698 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-operator-scripts\") pod \"heat-db-create-bzms5\" (UID: \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\") " pod="openstack/heat-db-create-bzms5" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.447769 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n5kp\" (UniqueName: \"kubernetes.io/projected/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-kube-api-access-5n5kp\") pod \"heat-db-create-bzms5\" (UID: \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\") " pod="openstack/heat-db-create-bzms5" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.461172 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkv2w\" (UniqueName: \"kubernetes.io/projected/80e46c2c-ff07-4431-ab74-fafdc1581673-kube-api-access-nkv2w\") pod \"barbican-5909-account-create-update-6pw92\" (UID: \"80e46c2c-ff07-4431-ab74-fafdc1581673\") " pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.467415 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.481173 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-a905-account-create-update-zr6j4"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.482295 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.485357 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.488573 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.493659 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a905-account-create-update-zr6j4"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.523165 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-operator-scripts\") pod \"cinder-db-create-xz6wd\" (UID: \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\") " pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.523456 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l29mb\" (UniqueName: \"kubernetes.io/projected/327b4198-11d6-4b61-829e-ef09616182b4-kube-api-access-l29mb\") pod \"heat-a905-account-create-update-zr6j4\" (UID: \"327b4198-11d6-4b61-829e-ef09616182b4\") " pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.523580 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/327b4198-11d6-4b61-829e-ef09616182b4-operator-scripts\") pod \"heat-a905-account-create-update-zr6j4\" (UID: \"327b4198-11d6-4b61-829e-ef09616182b4\") " pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.523687 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94kgw\" (UniqueName: \"kubernetes.io/projected/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-kube-api-access-94kgw\") pod \"cinder-db-create-xz6wd\" (UID: \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\") " pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.543419 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-rptl7"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.544498 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.547092 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.547378 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fmb5b" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.547983 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.548046 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.582550 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rptl7"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.608639 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bzms5" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.616194 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.646245 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-operator-scripts\") pod \"cinder-db-create-xz6wd\" (UID: \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\") " pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.646343 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l29mb\" (UniqueName: \"kubernetes.io/projected/327b4198-11d6-4b61-829e-ef09616182b4-kube-api-access-l29mb\") pod \"heat-a905-account-create-update-zr6j4\" (UID: \"327b4198-11d6-4b61-829e-ef09616182b4\") " pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.646406 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/327b4198-11d6-4b61-829e-ef09616182b4-operator-scripts\") pod \"heat-a905-account-create-update-zr6j4\" (UID: \"327b4198-11d6-4b61-829e-ef09616182b4\") " pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.646488 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94kgw\" (UniqueName: \"kubernetes.io/projected/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-kube-api-access-94kgw\") pod \"cinder-db-create-xz6wd\" (UID: \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\") " pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.647120 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-operator-scripts\") pod \"cinder-db-create-xz6wd\" (UID: \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\") " pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.648533 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/327b4198-11d6-4b61-829e-ef09616182b4-operator-scripts\") pod \"heat-a905-account-create-update-zr6j4\" (UID: \"327b4198-11d6-4b61-829e-ef09616182b4\") " pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.678170 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l29mb\" (UniqueName: \"kubernetes.io/projected/327b4198-11d6-4b61-829e-ef09616182b4-kube-api-access-l29mb\") pod \"heat-a905-account-create-update-zr6j4\" (UID: \"327b4198-11d6-4b61-829e-ef09616182b4\") " pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.696802 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8029-account-create-update-j2rvm"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.700459 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.755159 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-combined-ca-bundle\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.887299 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rf8x\" (UniqueName: \"kubernetes.io/projected/5ef68788-39b1-4011-bd01-a63d7520f540-kube-api-access-7rf8x\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.887468 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-config-data\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.888051 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.889390 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94kgw\" (UniqueName: \"kubernetes.io/projected/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-kube-api-access-94kgw\") pod \"cinder-db-create-xz6wd\" (UID: \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\") " pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.890457 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8029-account-create-update-j2rvm"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.899155 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9kk66"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.900346 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.917545 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9kk66"] Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.946303 4876 generic.go:334] "Generic (PLEG): container finished" podID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerID="3a63fd7d71efca11be3d44d274811be02f2758a0769977e828e60ba2761b9b43" exitCode=0 Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.946357 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" event={"ID":"c3dadbab-64b6-44e6-b07c-bc5372ffb775","Type":"ContainerDied","Data":"3a63fd7d71efca11be3d44d274811be02f2758a0769977e828e60ba2761b9b43"} Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.955952 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.989837 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9knr4\" (UniqueName: \"kubernetes.io/projected/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-kube-api-access-9knr4\") pod \"neutron-8029-account-create-update-j2rvm\" (UID: \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\") " pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.990261 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-operator-scripts\") pod \"neutron-8029-account-create-update-j2rvm\" (UID: \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\") " pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.990331 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-config-data\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.990410 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-combined-ca-bundle\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:28 crc kubenswrapper[4876]: I1205 08:35:28.990500 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rf8x\" (UniqueName: \"kubernetes.io/projected/5ef68788-39b1-4011-bd01-a63d7520f540-kube-api-access-7rf8x\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:28.997613 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-config-data\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.023186 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-combined-ca-bundle\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.024162 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rf8x\" (UniqueName: \"kubernetes.io/projected/5ef68788-39b1-4011-bd01-a63d7520f540-kube-api-access-7rf8x\") pod \"keystone-db-sync-rptl7\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.037277 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.096070 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9knr4\" (UniqueName: \"kubernetes.io/projected/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-kube-api-access-9knr4\") pod \"neutron-8029-account-create-update-j2rvm\" (UID: \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\") " pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.096190 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvjb7\" (UniqueName: \"kubernetes.io/projected/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-kube-api-access-wvjb7\") pod \"neutron-db-create-9kk66\" (UID: \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\") " pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.096249 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-operator-scripts\") pod \"neutron-db-create-9kk66\" (UID: \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\") " pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.096287 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-operator-scripts\") pod \"neutron-8029-account-create-update-j2rvm\" (UID: \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\") " pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.097218 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-operator-scripts\") pod \"neutron-8029-account-create-update-j2rvm\" (UID: \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\") " pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.119054 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9knr4\" (UniqueName: \"kubernetes.io/projected/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-kube-api-access-9knr4\") pod \"neutron-8029-account-create-update-j2rvm\" (UID: \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\") " pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.197446 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5909-account-create-update-6pw92"] Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.197734 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvjb7\" (UniqueName: \"kubernetes.io/projected/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-kube-api-access-wvjb7\") pod \"neutron-db-create-9kk66\" (UID: \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\") " pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.197803 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-operator-scripts\") pod \"neutron-db-create-9kk66\" (UID: \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\") " pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.198932 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-operator-scripts\") pod \"neutron-db-create-9kk66\" (UID: \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\") " pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.204592 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.223702 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvjb7\" (UniqueName: \"kubernetes.io/projected/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-kube-api-access-wvjb7\") pod \"neutron-db-create-9kk66\" (UID: \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\") " pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.261476 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-bzms5"] Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.270383 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6e8e-account-create-update-jnkdv"] Dec 05 08:35:29 crc kubenswrapper[4876]: W1205 08:35:29.273979 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9855cf16_5f53_4795_9103_3ee2c6b75f79.slice/crio-b4e8ad90ab3428d198a9b759df6245b1a09a6392d629bee234396bf30f12bcf2 WatchSource:0}: Error finding container b4e8ad90ab3428d198a9b759df6245b1a09a6392d629bee234396bf30f12bcf2: Status 404 returned error can't find the container with id b4e8ad90ab3428d198a9b759df6245b1a09a6392d629bee234396bf30f12bcf2 Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.276356 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.390079 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a905-account-create-update-zr6j4"] Dec 05 08:35:29 crc kubenswrapper[4876]: W1205 08:35:29.412999 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod327b4198_11d6_4b61_829e_ef09616182b4.slice/crio-611c51bc457ece11957d37cdc2b333ec1e9e539f4c7f338a2e08d87b339a8374 WatchSource:0}: Error finding container 611c51bc457ece11957d37cdc2b333ec1e9e539f4c7f338a2e08d87b339a8374: Status 404 returned error can't find the container with id 611c51bc457ece11957d37cdc2b333ec1e9e539f4c7f338a2e08d87b339a8374 Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.494982 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-mm4br"] Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.519766 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.646497 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-xz6wd"] Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.768888 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.828941 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-sb\") pod \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.829072 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxbx4\" (UniqueName: \"kubernetes.io/projected/c3dadbab-64b6-44e6-b07c-bc5372ffb775-kube-api-access-gxbx4\") pod \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.829152 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-config\") pod \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.829177 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-nb\") pod \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.829268 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-dns-svc\") pod \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\" (UID: \"c3dadbab-64b6-44e6-b07c-bc5372ffb775\") " Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.864056 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8029-account-create-update-j2rvm"] Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.892829 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3dadbab-64b6-44e6-b07c-bc5372ffb775-kube-api-access-gxbx4" (OuterVolumeSpecName: "kube-api-access-gxbx4") pod "c3dadbab-64b6-44e6-b07c-bc5372ffb775" (UID: "c3dadbab-64b6-44e6-b07c-bc5372ffb775"). InnerVolumeSpecName "kube-api-access-gxbx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.898484 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-rptl7"] Dec 05 08:35:29 crc kubenswrapper[4876]: W1205 08:35:29.917927 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ef68788_39b1_4011_bd01_a63d7520f540.slice/crio-7a3911b5c1c22468a3264d4c72c7d0350e7f7c64003d1d3f0a8b074cb39ccafd WatchSource:0}: Error finding container 7a3911b5c1c22468a3264d4c72c7d0350e7f7c64003d1d3f0a8b074cb39ccafd: Status 404 returned error can't find the container with id 7a3911b5c1c22468a3264d4c72c7d0350e7f7c64003d1d3f0a8b074cb39ccafd Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.932709 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxbx4\" (UniqueName: \"kubernetes.io/projected/c3dadbab-64b6-44e6-b07c-bc5372ffb775-kube-api-access-gxbx4\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.956379 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mm4br" event={"ID":"db2de57d-f90f-44c1-89de-d7706bd6e64c","Type":"ContainerStarted","Data":"d0ff2e1bb50ff2155db58f2878214cb844b4a3c3988acd19ca8002badb848be5"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.958079 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5909-account-create-update-6pw92" event={"ID":"80e46c2c-ff07-4431-ab74-fafdc1581673","Type":"ContainerStarted","Data":"1f3cd03084dd53a8b8bd3a8bfd434e10b31f04509515d7e6fc32aed83d65e77a"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.959778 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6e8e-account-create-update-jnkdv" event={"ID":"9855cf16-5f53-4795-9103-3ee2c6b75f79","Type":"ContainerStarted","Data":"b4e8ad90ab3428d198a9b759df6245b1a09a6392d629bee234396bf30f12bcf2"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.960843 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8029-account-create-update-j2rvm" event={"ID":"dabfca06-e3fb-4f63-baaf-f6b24bd4460a","Type":"ContainerStarted","Data":"5a79dab17b5e7e3a993615e6508ad30a730787b64dd0884460ce61898a8b0255"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.963093 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" event={"ID":"c3dadbab-64b6-44e6-b07c-bc5372ffb775","Type":"ContainerDied","Data":"e1bf612f2cdc05e67bbbc71bfb3c70c231487c295967e6481844a853eb93658f"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.963114 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gvpkj" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.963132 4876 scope.go:117] "RemoveContainer" containerID="3a63fd7d71efca11be3d44d274811be02f2758a0769977e828e60ba2761b9b43" Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.964015 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bzms5" event={"ID":"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f","Type":"ContainerStarted","Data":"aa459c8470e85b6eaa02c0f302b3b96638260ced2afb1596b21699f1d85f3d31"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.966462 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rptl7" event={"ID":"5ef68788-39b1-4011-bd01-a63d7520f540","Type":"ContainerStarted","Data":"7a3911b5c1c22468a3264d4c72c7d0350e7f7c64003d1d3f0a8b074cb39ccafd"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.967940 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a905-account-create-update-zr6j4" event={"ID":"327b4198-11d6-4b61-829e-ef09616182b4","Type":"ContainerStarted","Data":"611c51bc457ece11957d37cdc2b333ec1e9e539f4c7f338a2e08d87b339a8374"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.969218 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xz6wd" event={"ID":"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be","Type":"ContainerStarted","Data":"1a2d2c29767e82e2876b8d48c46541722c42db76c91bf2724b40e4fa2e14d276"} Dec 05 08:35:29 crc kubenswrapper[4876]: I1205 08:35:29.980174 4876 scope.go:117] "RemoveContainer" containerID="5636413e75e741732c3d3120266d0e141d7608ff5bb6ae372a17715d70b24e75" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.028934 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c3dadbab-64b6-44e6-b07c-bc5372ffb775" (UID: "c3dadbab-64b6-44e6-b07c-bc5372ffb775"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.031430 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c3dadbab-64b6-44e6-b07c-bc5372ffb775" (UID: "c3dadbab-64b6-44e6-b07c-bc5372ffb775"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.034212 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.034240 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.037425 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-config" (OuterVolumeSpecName: "config") pod "c3dadbab-64b6-44e6-b07c-bc5372ffb775" (UID: "c3dadbab-64b6-44e6-b07c-bc5372ffb775"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.058542 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c3dadbab-64b6-44e6-b07c-bc5372ffb775" (UID: "c3dadbab-64b6-44e6-b07c-bc5372ffb775"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.062009 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9kk66"] Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.135965 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.136053 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c3dadbab-64b6-44e6-b07c-bc5372ffb775-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.295111 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gvpkj"] Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.302559 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gvpkj"] Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.979563 4876 generic.go:334] "Generic (PLEG): container finished" podID="db2de57d-f90f-44c1-89de-d7706bd6e64c" containerID="7905d76b31b558ffe5a19de47f7f4ff57d1727f04a59c68bae10b5b238ff26b2" exitCode=0 Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.979642 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mm4br" event={"ID":"db2de57d-f90f-44c1-89de-d7706bd6e64c","Type":"ContainerDied","Data":"7905d76b31b558ffe5a19de47f7f4ff57d1727f04a59c68bae10b5b238ff26b2"} Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.982141 4876 generic.go:334] "Generic (PLEG): container finished" podID="327b4198-11d6-4b61-829e-ef09616182b4" containerID="d206ea9f116f0d29a176869b29933ae54243c8aaa83d42034b916c1d98b04679" exitCode=0 Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.982186 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a905-account-create-update-zr6j4" event={"ID":"327b4198-11d6-4b61-829e-ef09616182b4","Type":"ContainerDied","Data":"d206ea9f116f0d29a176869b29933ae54243c8aaa83d42034b916c1d98b04679"} Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.984545 4876 generic.go:334] "Generic (PLEG): container finished" podID="1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f" containerID="bcea7a3dc3026cc23d48072974b77b7de8dd3d3b624c940ed7f0d2cf2e64666f" exitCode=0 Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.984596 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bzms5" event={"ID":"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f","Type":"ContainerDied","Data":"bcea7a3dc3026cc23d48072974b77b7de8dd3d3b624c940ed7f0d2cf2e64666f"} Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.986502 4876 generic.go:334] "Generic (PLEG): container finished" podID="08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be" containerID="2114de6fdf9833669bc77e6056c62bdef51b47d93782be620403cba1524f562b" exitCode=0 Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.986619 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xz6wd" event={"ID":"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be","Type":"ContainerDied","Data":"2114de6fdf9833669bc77e6056c62bdef51b47d93782be620403cba1524f562b"} Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.988958 4876 generic.go:334] "Generic (PLEG): container finished" podID="bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6" containerID="f33eb3e638dd9f48068de5845629722c4abfbfb5789913f149f65ca34b79cb20" exitCode=0 Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.989016 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9kk66" event={"ID":"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6","Type":"ContainerDied","Data":"f33eb3e638dd9f48068de5845629722c4abfbfb5789913f149f65ca34b79cb20"} Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.989073 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9kk66" event={"ID":"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6","Type":"ContainerStarted","Data":"ea130fa865e4040da650c70aefde2660a2d5cadec8926497ac2e3564b7209a61"} Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.998184 4876 generic.go:334] "Generic (PLEG): container finished" podID="80e46c2c-ff07-4431-ab74-fafdc1581673" containerID="2f8e03026f257a88664deb69ca0e4180b7c2c4fd986841ab7e9c921ac63b1174" exitCode=0 Dec 05 08:35:30 crc kubenswrapper[4876]: I1205 08:35:30.998288 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5909-account-create-update-6pw92" event={"ID":"80e46c2c-ff07-4431-ab74-fafdc1581673","Type":"ContainerDied","Data":"2f8e03026f257a88664deb69ca0e4180b7c2c4fd986841ab7e9c921ac63b1174"} Dec 05 08:35:31 crc kubenswrapper[4876]: I1205 08:35:31.000829 4876 generic.go:334] "Generic (PLEG): container finished" podID="9855cf16-5f53-4795-9103-3ee2c6b75f79" containerID="7f5519a23a220ed25975e777bd47418dc954cbbbf9d4757dc699d7babaf9ac20" exitCode=0 Dec 05 08:35:31 crc kubenswrapper[4876]: I1205 08:35:31.000927 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6e8e-account-create-update-jnkdv" event={"ID":"9855cf16-5f53-4795-9103-3ee2c6b75f79","Type":"ContainerDied","Data":"7f5519a23a220ed25975e777bd47418dc954cbbbf9d4757dc699d7babaf9ac20"} Dec 05 08:35:31 crc kubenswrapper[4876]: I1205 08:35:31.002885 4876 generic.go:334] "Generic (PLEG): container finished" podID="dabfca06-e3fb-4f63-baaf-f6b24bd4460a" containerID="7b3ddfe2c86d27790ba0d10e5040d810ae8e8bd436c5fd8db96f31dfbc715646" exitCode=0 Dec 05 08:35:31 crc kubenswrapper[4876]: I1205 08:35:31.002961 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8029-account-create-update-j2rvm" event={"ID":"dabfca06-e3fb-4f63-baaf-f6b24bd4460a","Type":"ContainerDied","Data":"7b3ddfe2c86d27790ba0d10e5040d810ae8e8bd436c5fd8db96f31dfbc715646"} Dec 05 08:35:31 crc kubenswrapper[4876]: I1205 08:35:31.836116 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" path="/var/lib/kubelet/pods/c3dadbab-64b6-44e6-b07c-bc5372ffb775/volumes" Dec 05 08:35:33 crc kubenswrapper[4876]: I1205 08:35:33.024721 4876 generic.go:334] "Generic (PLEG): container finished" podID="2c58acae-5113-4e33-951e-409855923e87" containerID="bfe4c276c30394787954befcccf806035dc8eef3b38dddf3dc8b4e446b5807c4" exitCode=0 Dec 05 08:35:33 crc kubenswrapper[4876]: I1205 08:35:33.024811 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n7phf" event={"ID":"2c58acae-5113-4e33-951e-409855923e87","Type":"ContainerDied","Data":"bfe4c276c30394787954befcccf806035dc8eef3b38dddf3dc8b4e446b5807c4"} Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.324312 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.360593 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bzms5" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.368398 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.379819 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.394416 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.397878 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e46c2c-ff07-4431-ab74-fafdc1581673-operator-scripts\") pod \"80e46c2c-ff07-4431-ab74-fafdc1581673\" (UID: \"80e46c2c-ff07-4431-ab74-fafdc1581673\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.397951 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db2de57d-f90f-44c1-89de-d7706bd6e64c-operator-scripts\") pod \"db2de57d-f90f-44c1-89de-d7706bd6e64c\" (UID: \"db2de57d-f90f-44c1-89de-d7706bd6e64c\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.397975 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9knr4\" (UniqueName: \"kubernetes.io/projected/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-kube-api-access-9knr4\") pod \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\" (UID: \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.397995 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-operator-scripts\") pod \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\" (UID: \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.398013 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkv2w\" (UniqueName: \"kubernetes.io/projected/80e46c2c-ff07-4431-ab74-fafdc1581673-kube-api-access-nkv2w\") pod \"80e46c2c-ff07-4431-ab74-fafdc1581673\" (UID: \"80e46c2c-ff07-4431-ab74-fafdc1581673\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.398034 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n5kp\" (UniqueName: \"kubernetes.io/projected/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-kube-api-access-5n5kp\") pod \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\" (UID: \"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.398078 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-operator-scripts\") pod \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\" (UID: \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.398091 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-operator-scripts\") pod \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\" (UID: \"dabfca06-e3fb-4f63-baaf-f6b24bd4460a\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.398129 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwvgm\" (UniqueName: \"kubernetes.io/projected/db2de57d-f90f-44c1-89de-d7706bd6e64c-kube-api-access-nwvgm\") pod \"db2de57d-f90f-44c1-89de-d7706bd6e64c\" (UID: \"db2de57d-f90f-44c1-89de-d7706bd6e64c\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.398149 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94kgw\" (UniqueName: \"kubernetes.io/projected/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-kube-api-access-94kgw\") pod \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\" (UID: \"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.399514 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be" (UID: "08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.400724 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80e46c2c-ff07-4431-ab74-fafdc1581673-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "80e46c2c-ff07-4431-ab74-fafdc1581673" (UID: "80e46c2c-ff07-4431-ab74-fafdc1581673"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.400804 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f" (UID: "1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.401192 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dabfca06-e3fb-4f63-baaf-f6b24bd4460a" (UID: "dabfca06-e3fb-4f63-baaf-f6b24bd4460a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.403279 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-kube-api-access-5n5kp" (OuterVolumeSpecName: "kube-api-access-5n5kp") pod "1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f" (UID: "1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f"). InnerVolumeSpecName "kube-api-access-5n5kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.403670 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db2de57d-f90f-44c1-89de-d7706bd6e64c-kube-api-access-nwvgm" (OuterVolumeSpecName: "kube-api-access-nwvgm") pod "db2de57d-f90f-44c1-89de-d7706bd6e64c" (UID: "db2de57d-f90f-44c1-89de-d7706bd6e64c"). InnerVolumeSpecName "kube-api-access-nwvgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.404529 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db2de57d-f90f-44c1-89de-d7706bd6e64c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "db2de57d-f90f-44c1-89de-d7706bd6e64c" (UID: "db2de57d-f90f-44c1-89de-d7706bd6e64c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.408935 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-kube-api-access-9knr4" (OuterVolumeSpecName: "kube-api-access-9knr4") pod "dabfca06-e3fb-4f63-baaf-f6b24bd4460a" (UID: "dabfca06-e3fb-4f63-baaf-f6b24bd4460a"). InnerVolumeSpecName "kube-api-access-9knr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.411912 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-kube-api-access-94kgw" (OuterVolumeSpecName: "kube-api-access-94kgw") pod "08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be" (UID: "08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be"). InnerVolumeSpecName "kube-api-access-94kgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.424619 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e46c2c-ff07-4431-ab74-fafdc1581673-kube-api-access-nkv2w" (OuterVolumeSpecName: "kube-api-access-nkv2w") pod "80e46c2c-ff07-4431-ab74-fafdc1581673" (UID: "80e46c2c-ff07-4431-ab74-fafdc1581673"). InnerVolumeSpecName "kube-api-access-nkv2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.458595 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.465149 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.484302 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.487919 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.499746 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-config-data\") pod \"2c58acae-5113-4e33-951e-409855923e87\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.499808 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-676pp\" (UniqueName: \"kubernetes.io/projected/9855cf16-5f53-4795-9103-3ee2c6b75f79-kube-api-access-676pp\") pod \"9855cf16-5f53-4795-9103-3ee2c6b75f79\" (UID: \"9855cf16-5f53-4795-9103-3ee2c6b75f79\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.499869 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-operator-scripts\") pod \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\" (UID: \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.500552 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6" (UID: "bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501112 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/db2de57d-f90f-44c1-89de-d7706bd6e64c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501219 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9knr4\" (UniqueName: \"kubernetes.io/projected/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-kube-api-access-9knr4\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501284 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501301 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkv2w\" (UniqueName: \"kubernetes.io/projected/80e46c2c-ff07-4431-ab74-fafdc1581673-kube-api-access-nkv2w\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501317 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n5kp\" (UniqueName: \"kubernetes.io/projected/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f-kube-api-access-5n5kp\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501356 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501369 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dabfca06-e3fb-4f63-baaf-f6b24bd4460a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501382 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwvgm\" (UniqueName: \"kubernetes.io/projected/db2de57d-f90f-44c1-89de-d7706bd6e64c-kube-api-access-nwvgm\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501395 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94kgw\" (UniqueName: \"kubernetes.io/projected/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be-kube-api-access-94kgw\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501430 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.501470 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/80e46c2c-ff07-4431-ab74-fafdc1581673-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.510226 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9855cf16-5f53-4795-9103-3ee2c6b75f79-kube-api-access-676pp" (OuterVolumeSpecName: "kube-api-access-676pp") pod "9855cf16-5f53-4795-9103-3ee2c6b75f79" (UID: "9855cf16-5f53-4795-9103-3ee2c6b75f79"). InnerVolumeSpecName "kube-api-access-676pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.555577 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-config-data" (OuterVolumeSpecName: "config-data") pod "2c58acae-5113-4e33-951e-409855923e87" (UID: "2c58acae-5113-4e33-951e-409855923e87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.602159 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9855cf16-5f53-4795-9103-3ee2c6b75f79-operator-scripts\") pod \"9855cf16-5f53-4795-9103-3ee2c6b75f79\" (UID: \"9855cf16-5f53-4795-9103-3ee2c6b75f79\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.602250 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvjb7\" (UniqueName: \"kubernetes.io/projected/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-kube-api-access-wvjb7\") pod \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\" (UID: \"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.602286 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-combined-ca-bundle\") pod \"2c58acae-5113-4e33-951e-409855923e87\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.602402 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/327b4198-11d6-4b61-829e-ef09616182b4-operator-scripts\") pod \"327b4198-11d6-4b61-829e-ef09616182b4\" (UID: \"327b4198-11d6-4b61-829e-ef09616182b4\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.602471 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-db-sync-config-data\") pod \"2c58acae-5113-4e33-951e-409855923e87\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.602545 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d92bv\" (UniqueName: \"kubernetes.io/projected/2c58acae-5113-4e33-951e-409855923e87-kube-api-access-d92bv\") pod \"2c58acae-5113-4e33-951e-409855923e87\" (UID: \"2c58acae-5113-4e33-951e-409855923e87\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.602587 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l29mb\" (UniqueName: \"kubernetes.io/projected/327b4198-11d6-4b61-829e-ef09616182b4-kube-api-access-l29mb\") pod \"327b4198-11d6-4b61-829e-ef09616182b4\" (UID: \"327b4198-11d6-4b61-829e-ef09616182b4\") " Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.602917 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/327b4198-11d6-4b61-829e-ef09616182b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "327b4198-11d6-4b61-829e-ef09616182b4" (UID: "327b4198-11d6-4b61-829e-ef09616182b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.603151 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/327b4198-11d6-4b61-829e-ef09616182b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.603176 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.603300 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-676pp\" (UniqueName: \"kubernetes.io/projected/9855cf16-5f53-4795-9103-3ee2c6b75f79-kube-api-access-676pp\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.603445 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9855cf16-5f53-4795-9103-3ee2c6b75f79-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9855cf16-5f53-4795-9103-3ee2c6b75f79" (UID: "9855cf16-5f53-4795-9103-3ee2c6b75f79"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.605560 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-kube-api-access-wvjb7" (OuterVolumeSpecName: "kube-api-access-wvjb7") pod "bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6" (UID: "bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6"). InnerVolumeSpecName "kube-api-access-wvjb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.606141 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/327b4198-11d6-4b61-829e-ef09616182b4-kube-api-access-l29mb" (OuterVolumeSpecName: "kube-api-access-l29mb") pod "327b4198-11d6-4b61-829e-ef09616182b4" (UID: "327b4198-11d6-4b61-829e-ef09616182b4"). InnerVolumeSpecName "kube-api-access-l29mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.606278 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c58acae-5113-4e33-951e-409855923e87-kube-api-access-d92bv" (OuterVolumeSpecName: "kube-api-access-d92bv") pod "2c58acae-5113-4e33-951e-409855923e87" (UID: "2c58acae-5113-4e33-951e-409855923e87"). InnerVolumeSpecName "kube-api-access-d92bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.606576 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2c58acae-5113-4e33-951e-409855923e87" (UID: "2c58acae-5113-4e33-951e-409855923e87"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.622109 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c58acae-5113-4e33-951e-409855923e87" (UID: "2c58acae-5113-4e33-951e-409855923e87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.703849 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9855cf16-5f53-4795-9103-3ee2c6b75f79-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.703874 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvjb7\" (UniqueName: \"kubernetes.io/projected/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6-kube-api-access-wvjb7\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.703885 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.703904 4876 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2c58acae-5113-4e33-951e-409855923e87-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.703913 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l29mb\" (UniqueName: \"kubernetes.io/projected/327b4198-11d6-4b61-829e-ef09616182b4-kube-api-access-l29mb\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:34 crc kubenswrapper[4876]: I1205 08:35:34.703921 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d92bv\" (UniqueName: \"kubernetes.io/projected/2c58acae-5113-4e33-951e-409855923e87-kube-api-access-d92bv\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.057293 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8029-account-create-update-j2rvm" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.057500 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8029-account-create-update-j2rvm" event={"ID":"dabfca06-e3fb-4f63-baaf-f6b24bd4460a","Type":"ContainerDied","Data":"5a79dab17b5e7e3a993615e6508ad30a730787b64dd0884460ce61898a8b0255"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.057560 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a79dab17b5e7e3a993615e6508ad30a730787b64dd0884460ce61898a8b0255" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.066323 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bzms5" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.066342 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bzms5" event={"ID":"1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f","Type":"ContainerDied","Data":"aa459c8470e85b6eaa02c0f302b3b96638260ced2afb1596b21699f1d85f3d31"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.066394 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa459c8470e85b6eaa02c0f302b3b96638260ced2afb1596b21699f1d85f3d31" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.081091 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9kk66" event={"ID":"bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6","Type":"ContainerDied","Data":"ea130fa865e4040da650c70aefde2660a2d5cadec8926497ac2e3564b7209a61"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.081246 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea130fa865e4040da650c70aefde2660a2d5cadec8926497ac2e3564b7209a61" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.081481 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9kk66" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.106182 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6e8e-account-create-update-jnkdv" event={"ID":"9855cf16-5f53-4795-9103-3ee2c6b75f79","Type":"ContainerDied","Data":"b4e8ad90ab3428d198a9b759df6245b1a09a6392d629bee234396bf30f12bcf2"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.106249 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4e8ad90ab3428d198a9b759df6245b1a09a6392d629bee234396bf30f12bcf2" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.106354 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6e8e-account-create-update-jnkdv" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.113456 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-n7phf" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.113472 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-n7phf" event={"ID":"2c58acae-5113-4e33-951e-409855923e87","Type":"ContainerDied","Data":"19da0bd05f039a884a5201c7d5280ddf3345d22fa708081943e11272e74377e0"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.114309 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19da0bd05f039a884a5201c7d5280ddf3345d22fa708081943e11272e74377e0" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.117766 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a905-account-create-update-zr6j4" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.117758 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a905-account-create-update-zr6j4" event={"ID":"327b4198-11d6-4b61-829e-ef09616182b4","Type":"ContainerDied","Data":"611c51bc457ece11957d37cdc2b333ec1e9e539f4c7f338a2e08d87b339a8374"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.117964 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="611c51bc457ece11957d37cdc2b333ec1e9e539f4c7f338a2e08d87b339a8374" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.121527 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-xz6wd" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.121519 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-xz6wd" event={"ID":"08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be","Type":"ContainerDied","Data":"1a2d2c29767e82e2876b8d48c46541722c42db76c91bf2724b40e4fa2e14d276"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.121761 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a2d2c29767e82e2876b8d48c46541722c42db76c91bf2724b40e4fa2e14d276" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.126126 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rptl7" event={"ID":"5ef68788-39b1-4011-bd01-a63d7520f540","Type":"ContainerStarted","Data":"c0c041a52c79ac54e83bfddc4d7fea542eb2845f3ca1890eeedb1ebd3a60ab32"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.131004 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-mm4br" event={"ID":"db2de57d-f90f-44c1-89de-d7706bd6e64c","Type":"ContainerDied","Data":"d0ff2e1bb50ff2155db58f2878214cb844b4a3c3988acd19ca8002badb848be5"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.131054 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0ff2e1bb50ff2155db58f2878214cb844b4a3c3988acd19ca8002badb848be5" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.131134 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-mm4br" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.135774 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5909-account-create-update-6pw92" event={"ID":"80e46c2c-ff07-4431-ab74-fafdc1581673","Type":"ContainerDied","Data":"1f3cd03084dd53a8b8bd3a8bfd434e10b31f04509515d7e6fc32aed83d65e77a"} Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.135813 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f3cd03084dd53a8b8bd3a8bfd434e10b31f04509515d7e6fc32aed83d65e77a" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.135933 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5909-account-create-update-6pw92" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.164874 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-rptl7" podStartSLOduration=2.914466446 podStartE2EDuration="7.164851927s" podCreationTimestamp="2025-12-05 08:35:28 +0000 UTC" firstStartedPulling="2025-12-05 08:35:29.920505361 +0000 UTC m=+1194.409169983" lastFinishedPulling="2025-12-05 08:35:34.170890842 +0000 UTC m=+1198.659555464" observedRunningTime="2025-12-05 08:35:35.154131474 +0000 UTC m=+1199.642796096" watchObservedRunningTime="2025-12-05 08:35:35.164851927 +0000 UTC m=+1199.653516559" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.526952 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-lvk22"] Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527351 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527387 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527399 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dabfca06-e3fb-4f63-baaf-f6b24bd4460a" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527410 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="dabfca06-e3fb-4f63-baaf-f6b24bd4460a" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527436 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327b4198-11d6-4b61-829e-ef09616182b4" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527446 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="327b4198-11d6-4b61-829e-ef09616182b4" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527467 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db2de57d-f90f-44c1-89de-d7706bd6e64c" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527475 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="db2de57d-f90f-44c1-89de-d7706bd6e64c" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527489 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e46c2c-ff07-4431-ab74-fafdc1581673" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527497 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e46c2c-ff07-4431-ab74-fafdc1581673" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527508 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerName="init" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527517 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerName="init" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527528 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527535 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527548 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c58acae-5113-4e33-951e-409855923e87" containerName="glance-db-sync" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527556 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c58acae-5113-4e33-951e-409855923e87" containerName="glance-db-sync" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527569 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9855cf16-5f53-4795-9103-3ee2c6b75f79" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527577 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9855cf16-5f53-4795-9103-3ee2c6b75f79" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527598 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerName="dnsmasq-dns" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527607 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerName="dnsmasq-dns" Dec 05 08:35:35 crc kubenswrapper[4876]: E1205 08:35:35.527618 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527635 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527847 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e46c2c-ff07-4431-ab74-fafdc1581673" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527870 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527885 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="db2de57d-f90f-44c1-89de-d7706bd6e64c" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527931 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9855cf16-5f53-4795-9103-3ee2c6b75f79" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527947 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527962 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be" containerName="mariadb-database-create" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527973 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="327b4198-11d6-4b61-829e-ef09616182b4" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527985 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="dabfca06-e3fb-4f63-baaf-f6b24bd4460a" containerName="mariadb-account-create-update" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.527999 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c58acae-5113-4e33-951e-409855923e87" containerName="glance-db-sync" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.528008 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3dadbab-64b6-44e6-b07c-bc5372ffb775" containerName="dnsmasq-dns" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.529048 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.585275 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-lvk22"] Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.627412 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-svc\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.627503 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.627542 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-config\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.627561 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.627688 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.627769 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzvc8\" (UniqueName: \"kubernetes.io/projected/d7d2d7df-4d7b-4417-8405-a358a6ff4961-kube-api-access-pzvc8\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.729391 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-svc\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.729437 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.729475 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-config\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.729502 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.729580 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.729609 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzvc8\" (UniqueName: \"kubernetes.io/projected/d7d2d7df-4d7b-4417-8405-a358a6ff4961-kube-api-access-pzvc8\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.730457 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.730663 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-config\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.732248 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-svc\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.732445 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.732485 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.749240 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzvc8\" (UniqueName: \"kubernetes.io/projected/d7d2d7df-4d7b-4417-8405-a358a6ff4961-kube-api-access-pzvc8\") pod \"dnsmasq-dns-895cf5cf-lvk22\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:35 crc kubenswrapper[4876]: I1205 08:35:35.849739 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:36 crc kubenswrapper[4876]: I1205 08:35:36.277701 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-lvk22"] Dec 05 08:35:36 crc kubenswrapper[4876]: W1205 08:35:36.279672 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7d2d7df_4d7b_4417_8405_a358a6ff4961.slice/crio-ddea0a63cd1e0b64b486b430326ab91bf3c7b3c67dd7cda3d4d2cf9590543f6d WatchSource:0}: Error finding container ddea0a63cd1e0b64b486b430326ab91bf3c7b3c67dd7cda3d4d2cf9590543f6d: Status 404 returned error can't find the container with id ddea0a63cd1e0b64b486b430326ab91bf3c7b3c67dd7cda3d4d2cf9590543f6d Dec 05 08:35:37 crc kubenswrapper[4876]: I1205 08:35:37.156352 4876 generic.go:334] "Generic (PLEG): container finished" podID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" containerID="95bef110b7c57caa88e41c1d5bb1a356d9088b5021efd97838c62ea2d79f78d7" exitCode=0 Dec 05 08:35:37 crc kubenswrapper[4876]: I1205 08:35:37.156495 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" event={"ID":"d7d2d7df-4d7b-4417-8405-a358a6ff4961","Type":"ContainerDied","Data":"95bef110b7c57caa88e41c1d5bb1a356d9088b5021efd97838c62ea2d79f78d7"} Dec 05 08:35:37 crc kubenswrapper[4876]: I1205 08:35:37.156738 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" event={"ID":"d7d2d7df-4d7b-4417-8405-a358a6ff4961","Type":"ContainerStarted","Data":"ddea0a63cd1e0b64b486b430326ab91bf3c7b3c67dd7cda3d4d2cf9590543f6d"} Dec 05 08:35:38 crc kubenswrapper[4876]: I1205 08:35:38.170145 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" event={"ID":"d7d2d7df-4d7b-4417-8405-a358a6ff4961","Type":"ContainerStarted","Data":"d155ac85e46f65bd157a53b0ea300e63cd3a5d13c94d8f0eb0cffdf5ef36e807"} Dec 05 08:35:38 crc kubenswrapper[4876]: I1205 08:35:38.170540 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:38 crc kubenswrapper[4876]: I1205 08:35:38.174137 4876 generic.go:334] "Generic (PLEG): container finished" podID="5ef68788-39b1-4011-bd01-a63d7520f540" containerID="c0c041a52c79ac54e83bfddc4d7fea542eb2845f3ca1890eeedb1ebd3a60ab32" exitCode=0 Dec 05 08:35:38 crc kubenswrapper[4876]: I1205 08:35:38.174211 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rptl7" event={"ID":"5ef68788-39b1-4011-bd01-a63d7520f540","Type":"ContainerDied","Data":"c0c041a52c79ac54e83bfddc4d7fea542eb2845f3ca1890eeedb1ebd3a60ab32"} Dec 05 08:35:38 crc kubenswrapper[4876]: I1205 08:35:38.192869 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" podStartSLOduration=3.192843327 podStartE2EDuration="3.192843327s" podCreationTimestamp="2025-12-05 08:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:38.189789081 +0000 UTC m=+1202.678453713" watchObservedRunningTime="2025-12-05 08:35:38.192843327 +0000 UTC m=+1202.681507959" Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.490142 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.600498 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-config-data\") pod \"5ef68788-39b1-4011-bd01-a63d7520f540\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.600603 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-combined-ca-bundle\") pod \"5ef68788-39b1-4011-bd01-a63d7520f540\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.600695 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rf8x\" (UniqueName: \"kubernetes.io/projected/5ef68788-39b1-4011-bd01-a63d7520f540-kube-api-access-7rf8x\") pod \"5ef68788-39b1-4011-bd01-a63d7520f540\" (UID: \"5ef68788-39b1-4011-bd01-a63d7520f540\") " Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.605843 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef68788-39b1-4011-bd01-a63d7520f540-kube-api-access-7rf8x" (OuterVolumeSpecName: "kube-api-access-7rf8x") pod "5ef68788-39b1-4011-bd01-a63d7520f540" (UID: "5ef68788-39b1-4011-bd01-a63d7520f540"). InnerVolumeSpecName "kube-api-access-7rf8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.624504 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ef68788-39b1-4011-bd01-a63d7520f540" (UID: "5ef68788-39b1-4011-bd01-a63d7520f540"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.656290 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-config-data" (OuterVolumeSpecName: "config-data") pod "5ef68788-39b1-4011-bd01-a63d7520f540" (UID: "5ef68788-39b1-4011-bd01-a63d7520f540"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.702785 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rf8x\" (UniqueName: \"kubernetes.io/projected/5ef68788-39b1-4011-bd01-a63d7520f540-kube-api-access-7rf8x\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.702823 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:39 crc kubenswrapper[4876]: I1205 08:35:39.702834 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef68788-39b1-4011-bd01-a63d7520f540-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.193504 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-rptl7" event={"ID":"5ef68788-39b1-4011-bd01-a63d7520f540","Type":"ContainerDied","Data":"7a3911b5c1c22468a3264d4c72c7d0350e7f7c64003d1d3f0a8b074cb39ccafd"} Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.193547 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a3911b5c1c22468a3264d4c72c7d0350e7f7c64003d1d3f0a8b074cb39ccafd" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.193549 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-rptl7" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.465710 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-lvk22"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.466039 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" podUID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" containerName="dnsmasq-dns" containerID="cri-o://d155ac85e46f65bd157a53b0ea300e63cd3a5d13c94d8f0eb0cffdf5ef36e807" gracePeriod=10 Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.498987 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-r4cl4"] Dec 05 08:35:40 crc kubenswrapper[4876]: E1205 08:35:40.499369 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef68788-39b1-4011-bd01-a63d7520f540" containerName="keystone-db-sync" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.499382 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef68788-39b1-4011-bd01-a63d7520f540" containerName="keystone-db-sync" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.499556 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef68788-39b1-4011-bd01-a63d7520f540" containerName="keystone-db-sync" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.500461 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.507716 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ph7r9"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.509255 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.511626 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.512004 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.512268 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fmb5b" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.513053 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.513346 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.523975 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-r4cl4"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.534375 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ph7r9"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.618460 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwbjh\" (UniqueName: \"kubernetes.io/projected/4c78505d-915e-4312-afff-a878017bff60-kube-api-access-zwbjh\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.618771 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6gpt\" (UniqueName: \"kubernetes.io/projected/58730925-f69e-43f4-a7db-d624537852d3-kube-api-access-d6gpt\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.618919 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.619075 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.619220 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-scripts\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.619344 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.619521 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-fernet-keys\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.619629 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-credential-keys\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.619754 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-config\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.620103 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-combined-ca-bundle\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.620228 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-config-data\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.620402 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.641686 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-xbjb2"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.643017 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.649854 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.655539 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-52ct9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.663848 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xbjb2"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.718799 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2rwzp"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.720114 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.723794 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-scripts\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724154 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724197 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-fernet-keys\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724215 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-credential-keys\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724236 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-config\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724257 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-combined-ca-bundle\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724273 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-config-data\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724305 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724322 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwbjh\" (UniqueName: \"kubernetes.io/projected/4c78505d-915e-4312-afff-a878017bff60-kube-api-access-zwbjh\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724351 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6gpt\" (UniqueName: \"kubernetes.io/projected/58730925-f69e-43f4-a7db-d624537852d3-kube-api-access-d6gpt\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724380 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.724404 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.725025 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.725118 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.725766 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-config\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.726065 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.727119 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.727378 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.727741 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.727937 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cd7mp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.736671 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2rwzp"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.737970 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-scripts\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.738491 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-config-data\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.739912 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-fernet-keys\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.740490 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-combined-ca-bundle\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.756622 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-credential-keys\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.761650 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwbjh\" (UniqueName: \"kubernetes.io/projected/4c78505d-915e-4312-afff-a878017bff60-kube-api-access-zwbjh\") pod \"keystone-bootstrap-ph7r9\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.791864 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vxcln"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.793809 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.810593 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.810813 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-hmxgp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.810970 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.822347 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6gpt\" (UniqueName: \"kubernetes.io/projected/58730925-f69e-43f4-a7db-d624537852d3-kube-api-access-d6gpt\") pod \"dnsmasq-dns-6c9c9f998c-r4cl4\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.861934 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-combined-ca-bundle\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.862082 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs9gh\" (UniqueName: \"kubernetes.io/projected/207c5e9c-418a-4c07-8e8a-bb79948d855d-kube-api-access-gs9gh\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.867827 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-config-data\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.869755 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-etc-machine-id\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.869812 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-combined-ca-bundle\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.869853 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-config\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.870103 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-scripts\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.870188 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nptt\" (UniqueName: \"kubernetes.io/projected/5a1ae752-b331-4d0f-8039-e5720312f674-kube-api-access-2nptt\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.897473 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vxcln"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.952777 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.959958 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-m2khz"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.961031 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.973429 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mlf96" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.973687 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.974215 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-m2khz"] Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.974666 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.975926 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-db-sync-config-data\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.975973 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-etc-machine-id\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.975994 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-combined-ca-bundle\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976014 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-config\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976036 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55btw\" (UniqueName: \"kubernetes.io/projected/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-kube-api-access-55btw\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976068 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-combined-ca-bundle\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976091 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-scripts\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976118 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nptt\" (UniqueName: \"kubernetes.io/projected/5a1ae752-b331-4d0f-8039-e5720312f674-kube-api-access-2nptt\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976135 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-config-data\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976177 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-combined-ca-bundle\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976219 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs9gh\" (UniqueName: \"kubernetes.io/projected/207c5e9c-418a-4c07-8e8a-bb79948d855d-kube-api-access-gs9gh\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:40 crc kubenswrapper[4876]: I1205 08:35:40.976249 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-config-data\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:40.983515 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-etc-machine-id\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:40.992201 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-config-data\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:40.992726 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-combined-ca-bundle\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:40.993346 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-config\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:40.995567 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-combined-ca-bundle\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:40.999721 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-scripts\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.009864 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs9gh\" (UniqueName: \"kubernetes.io/projected/207c5e9c-418a-4c07-8e8a-bb79948d855d-kube-api-access-gs9gh\") pod \"neutron-db-sync-2rwzp\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.013110 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-n5bmx"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.014222 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.019100 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.019142 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.019366 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nptt\" (UniqueName: \"kubernetes.io/projected/5a1ae752-b331-4d0f-8039-e5720312f674-kube-api-access-2nptt\") pod \"heat-db-sync-xbjb2\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.019408 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-djmxn" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.042361 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-n5bmx"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.079569 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-config-data\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.079998 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-db-sync-config-data\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.080034 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-db-sync-config-data\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.080056 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-combined-ca-bundle\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.080090 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55btw\" (UniqueName: \"kubernetes.io/projected/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-kube-api-access-55btw\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.080122 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnv8k\" (UniqueName: \"kubernetes.io/projected/2d18b829-e516-41de-989d-bea32b940409-kube-api-access-hnv8k\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.080140 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-combined-ca-bundle\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.087244 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-combined-ca-bundle\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.087501 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-config-data\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.089931 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-db-sync-config-data\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.096749 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-r4cl4"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.101014 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55btw\" (UniqueName: \"kubernetes.io/projected/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-kube-api-access-55btw\") pod \"cinder-db-sync-vxcln\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.113151 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.116969 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.120712 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.120819 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.122307 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-hrms7"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.124066 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.144631 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.181506 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-db-sync-config-data\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.181551 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-combined-ca-bundle\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.181572 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-combined-ca-bundle\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.181619 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7k8q\" (UniqueName: \"kubernetes.io/projected/460f93bd-e732-42c0-8e75-625f3346800c-kube-api-access-q7k8q\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.181641 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnv8k\" (UniqueName: \"kubernetes.io/projected/2d18b829-e516-41de-989d-bea32b940409-kube-api-access-hnv8k\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.181679 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-config-data\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.181695 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/460f93bd-e732-42c0-8e75-625f3346800c-logs\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.181732 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-scripts\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.183678 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-hrms7"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.186800 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-db-sync-config-data\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.187317 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-combined-ca-bundle\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.197246 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.213145 4876 generic.go:334] "Generic (PLEG): container finished" podID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" containerID="d155ac85e46f65bd157a53b0ea300e63cd3a5d13c94d8f0eb0cffdf5ef36e807" exitCode=0 Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.213192 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" event={"ID":"d7d2d7df-4d7b-4417-8405-a358a6ff4961","Type":"ContainerDied","Data":"d155ac85e46f65bd157a53b0ea300e63cd3a5d13c94d8f0eb0cffdf5ef36e807"} Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.216919 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnv8k\" (UniqueName: \"kubernetes.io/projected/2d18b829-e516-41de-989d-bea32b940409-kube-api-access-hnv8k\") pod \"barbican-db-sync-m2khz\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.229194 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vxcln" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.282929 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283271 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-scripts\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283290 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-log-httpd\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283323 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283342 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283370 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-config\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283388 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283407 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5qgf\" (UniqueName: \"kubernetes.io/projected/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-kube-api-access-m5qgf\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283430 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-combined-ca-bundle\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283448 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvf6d\" (UniqueName: \"kubernetes.io/projected/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-kube-api-access-wvf6d\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283487 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-config-data\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283511 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7k8q\" (UniqueName: \"kubernetes.io/projected/460f93bd-e732-42c0-8e75-625f3346800c-kube-api-access-q7k8q\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283535 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-run-httpd\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-config-data\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283601 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/460f93bd-e732-42c0-8e75-625f3346800c-logs\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283617 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-scripts\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283630 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.283647 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.284465 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/460f93bd-e732-42c0-8e75-625f3346800c-logs\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.287430 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-combined-ca-bundle\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.288731 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-config-data\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.290989 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-scripts\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.291328 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xbjb2" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.304148 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7k8q\" (UniqueName: \"kubernetes.io/projected/460f93bd-e732-42c0-8e75-625f3346800c-kube-api-access-q7k8q\") pod \"placement-db-sync-n5bmx\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.337554 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.370306 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-m2khz" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.384999 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385053 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-log-httpd\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385096 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385117 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385144 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-config\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385163 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385180 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5qgf\" (UniqueName: \"kubernetes.io/projected/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-kube-api-access-m5qgf\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385208 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvf6d\" (UniqueName: \"kubernetes.io/projected/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-kube-api-access-wvf6d\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385243 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-config-data\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385270 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-run-httpd\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385311 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-scripts\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385327 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385342 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.385845 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.386142 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.393407 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n5bmx" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.396384 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.396390 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-config\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.396575 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.396605 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-log-httpd\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.409866 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-run-httpd\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.411222 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-config-data\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.411956 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvf6d\" (UniqueName: \"kubernetes.io/projected/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-kube-api-access-wvf6d\") pod \"dnsmasq-dns-57c957c4ff-hrms7\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.414493 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-scripts\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.416514 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.416609 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5qgf\" (UniqueName: \"kubernetes.io/projected/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-kube-api-access-m5qgf\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.418195 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.475712 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.481700 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.486368 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzvc8\" (UniqueName: \"kubernetes.io/projected/d7d2d7df-4d7b-4417-8405-a358a6ff4961-kube-api-access-pzvc8\") pod \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.486425 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-swift-storage-0\") pod \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.486472 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-nb\") pod \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.486511 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-config\") pod \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.486549 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-svc\") pod \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.486764 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-sb\") pod \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\" (UID: \"d7d2d7df-4d7b-4417-8405-a358a6ff4961\") " Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.503295 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d2d7df-4d7b-4417-8405-a358a6ff4961-kube-api-access-pzvc8" (OuterVolumeSpecName: "kube-api-access-pzvc8") pod "d7d2d7df-4d7b-4417-8405-a358a6ff4961" (UID: "d7d2d7df-4d7b-4417-8405-a358a6ff4961"). InnerVolumeSpecName "kube-api-access-pzvc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.570743 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ph7r9"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.589868 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzvc8\" (UniqueName: \"kubernetes.io/projected/d7d2d7df-4d7b-4417-8405-a358a6ff4961-kube-api-access-pzvc8\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.591037 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7d2d7df-4d7b-4417-8405-a358a6ff4961" (UID: "d7d2d7df-4d7b-4417-8405-a358a6ff4961"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:41 crc kubenswrapper[4876]: W1205 08:35:41.603219 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c78505d_915e_4312_afff_a878017bff60.slice/crio-5f4312d0d5be88d9991a3c3b9971e25ce438331bc3387b2a3435238a92839fc2 WatchSource:0}: Error finding container 5f4312d0d5be88d9991a3c3b9971e25ce438331bc3387b2a3435238a92839fc2: Status 404 returned error can't find the container with id 5f4312d0d5be88d9991a3c3b9971e25ce438331bc3387b2a3435238a92839fc2 Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.616216 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7d2d7df-4d7b-4417-8405-a358a6ff4961" (UID: "d7d2d7df-4d7b-4417-8405-a358a6ff4961"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.622743 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d7d2d7df-4d7b-4417-8405-a358a6ff4961" (UID: "d7d2d7df-4d7b-4417-8405-a358a6ff4961"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.665232 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-config" (OuterVolumeSpecName: "config") pod "d7d2d7df-4d7b-4417-8405-a358a6ff4961" (UID: "d7d2d7df-4d7b-4417-8405-a358a6ff4961"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.681941 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-r4cl4"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.687078 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7d2d7df-4d7b-4417-8405-a358a6ff4961" (UID: "d7d2d7df-4d7b-4417-8405-a358a6ff4961"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.689202 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:35:41 crc kubenswrapper[4876]: E1205 08:35:41.692638 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" containerName="dnsmasq-dns" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.692685 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" containerName="dnsmasq-dns" Dec 05 08:35:41 crc kubenswrapper[4876]: E1205 08:35:41.692765 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" containerName="init" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.692775 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" containerName="init" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.693120 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" containerName="dnsmasq-dns" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.697176 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.697293 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.697440 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.697523 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.697601 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d2d7df-4d7b-4417-8405-a358a6ff4961-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.698103 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.698307 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.702362 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.702496 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.702580 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.702648 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5msdq" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.794013 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.795443 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.800747 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-config-data\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.800805 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-scripts\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.800835 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.800856 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.801306 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdbcw\" (UniqueName: \"kubernetes.io/projected/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-kube-api-access-hdbcw\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.801346 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.801406 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.801598 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-logs\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.808551 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.809096 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.877985 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902589 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902642 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdbcw\" (UniqueName: \"kubernetes.io/projected/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-kube-api-access-hdbcw\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902661 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902681 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902700 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-logs\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902721 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902816 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902841 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902863 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902883 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k8sb\" (UniqueName: \"kubernetes.io/projected/2de56934-652e-4571-93c2-2a504945cd26-kube-api-access-9k8sb\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902922 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902941 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-logs\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902962 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-config-data\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.902986 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-scripts\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.903019 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.903038 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.904074 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.924408 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.924964 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-logs\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.953193 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.953870 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdbcw\" (UniqueName: \"kubernetes.io/projected/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-kube-api-access-hdbcw\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.956958 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-config-data\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.962635 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2rwzp"] Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.965007 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-scripts\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:41 crc kubenswrapper[4876]: I1205 08:35:41.972376 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.006538 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.006587 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.006609 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.006638 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-logs\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.006735 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.006788 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.006817 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k8sb\" (UniqueName: \"kubernetes.io/projected/2de56934-652e-4571-93c2-2a504945cd26-kube-api-access-9k8sb\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.006851 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.008063 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-logs\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.013431 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.022515 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.028079 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.028657 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.028968 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.038529 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.045804 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.059276 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.067431 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k8sb\" (UniqueName: \"kubernetes.io/projected/2de56934-652e-4571-93c2-2a504945cd26-kube-api-access-9k8sb\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.085865 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.229029 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2rwzp" event={"ID":"207c5e9c-418a-4c07-8e8a-bb79948d855d","Type":"ContainerStarted","Data":"1e0aae61edfe35977d5420788905b8ae21c239343e76bc171506193a8bc75c9a"} Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.230270 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xbjb2"] Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.231851 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" event={"ID":"58730925-f69e-43f4-a7db-d624537852d3","Type":"ContainerStarted","Data":"a4254c2631628786986355b4a6c8f4e63e6e184ad96bb2380d852d86e162865d"} Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.249077 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" event={"ID":"d7d2d7df-4d7b-4417-8405-a358a6ff4961","Type":"ContainerDied","Data":"ddea0a63cd1e0b64b486b430326ab91bf3c7b3c67dd7cda3d4d2cf9590543f6d"} Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.249134 4876 scope.go:117] "RemoveContainer" containerID="d155ac85e46f65bd157a53b0ea300e63cd3a5d13c94d8f0eb0cffdf5ef36e807" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.249287 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-lvk22" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.261413 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vxcln"] Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.265725 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ph7r9" event={"ID":"4c78505d-915e-4312-afff-a878017bff60","Type":"ContainerStarted","Data":"5f4312d0d5be88d9991a3c3b9971e25ce438331bc3387b2a3435238a92839fc2"} Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.292804 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-lvk22"] Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.296706 4876 scope.go:117] "RemoveContainer" containerID="95bef110b7c57caa88e41c1d5bb1a356d9088b5021efd97838c62ea2d79f78d7" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.302203 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-lvk22"] Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.311114 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-m2khz"] Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.384921 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.461175 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-n5bmx"] Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.481805 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:35:42 crc kubenswrapper[4876]: W1205 08:35:42.505219 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee3139f2_0c55_464e_bdbf_eefcc4bb59c5.slice/crio-d8e3dbac459cdc2ae7796f6937bb3a55b8a27bf4ba24513b6215e35cad65a059 WatchSource:0}: Error finding container d8e3dbac459cdc2ae7796f6937bb3a55b8a27bf4ba24513b6215e35cad65a059: Status 404 returned error can't find the container with id d8e3dbac459cdc2ae7796f6937bb3a55b8a27bf4ba24513b6215e35cad65a059 Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.632735 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-hrms7"] Dec 05 08:35:42 crc kubenswrapper[4876]: I1205 08:35:42.861260 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.023355 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.244124 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.321749 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2de56934-652e-4571-93c2-2a504945cd26","Type":"ContainerStarted","Data":"1fed6a4d40221ef202049ce908d3be11c86733e601a414dca3fe77b7df788474"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.327980 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.331316 4876 generic.go:334] "Generic (PLEG): container finished" podID="58730925-f69e-43f4-a7db-d624537852d3" containerID="95ae93c8b0a297929c92e8affcc66a29369ad235557b580d300b8b970be76643" exitCode=0 Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.331556 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" event={"ID":"58730925-f69e-43f4-a7db-d624537852d3","Type":"ContainerDied","Data":"95ae93c8b0a297929c92e8affcc66a29369ad235557b580d300b8b970be76643"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.337716 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549","Type":"ContainerStarted","Data":"f37b0b4e77f6abba7ef0ebbdddf944b021fab7324eca36ff9546ad1892b68215"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.339415 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ph7r9" event={"ID":"4c78505d-915e-4312-afff-a878017bff60","Type":"ContainerStarted","Data":"bfa61edc4a079881d538371e415544a2e93a9c3a589d05c7f54e6c19f38443ae"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.343334 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-m2khz" event={"ID":"2d18b829-e516-41de-989d-bea32b940409","Type":"ContainerStarted","Data":"43edfbacf605443d0d0944225e3d6de9f7ea2f4b7fb9c285098cf479b068b686"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.343374 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.346000 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2rwzp" event={"ID":"207c5e9c-418a-4c07-8e8a-bb79948d855d","Type":"ContainerStarted","Data":"cbc720abbc55b77c1cbdf9bb522e82ac772b3b46fdfb8f291d1fb783bcaf5ad2"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.355802 4876 generic.go:334] "Generic (PLEG): container finished" podID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" containerID="1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d" exitCode=0 Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.357183 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" event={"ID":"fd5f73d3-2859-4e4d-a5d8-02acd110f11a","Type":"ContainerDied","Data":"1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.357255 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" event={"ID":"fd5f73d3-2859-4e4d-a5d8-02acd110f11a","Type":"ContainerStarted","Data":"8e0033ba63663d8341cbe3946db8b436b6e7ff35b314ec1b52fff083b7049224"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.387339 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xbjb2" event={"ID":"5a1ae752-b331-4d0f-8039-e5720312f674","Type":"ContainerStarted","Data":"de2e048dbe1393e79577008e7b06a819b08783e180454a2f426b1caa25bcee00"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.398123 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n5bmx" event={"ID":"460f93bd-e732-42c0-8e75-625f3346800c","Type":"ContainerStarted","Data":"0711a02349f25779b80825587a7c137737de21cdae5259c054ad030ce613802c"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.432946 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vxcln" event={"ID":"62f1d7bc-1a36-4fc0-a94d-72fcedda052e","Type":"ContainerStarted","Data":"2988e26a23f24514afda91af74167f2c6be3667fe48867158bb1228754988893"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.438536 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5","Type":"ContainerStarted","Data":"d8e3dbac459cdc2ae7796f6937bb3a55b8a27bf4ba24513b6215e35cad65a059"} Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.443126 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2rwzp" podStartSLOduration=3.443104211 podStartE2EDuration="3.443104211s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:43.422339094 +0000 UTC m=+1207.911003726" watchObservedRunningTime="2025-12-05 08:35:43.443104211 +0000 UTC m=+1207.931768833" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.467084 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ph7r9" podStartSLOduration=3.467058988 podStartE2EDuration="3.467058988s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:43.441695271 +0000 UTC m=+1207.930359893" watchObservedRunningTime="2025-12-05 08:35:43.467058988 +0000 UTC m=+1207.955723610" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.785336 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.844589 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d2d7df-4d7b-4417-8405-a358a6ff4961" path="/var/lib/kubelet/pods/d7d2d7df-4d7b-4417-8405-a358a6ff4961/volumes" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.846504 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-nb\") pod \"58730925-f69e-43f4-a7db-d624537852d3\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.846548 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-sb\") pod \"58730925-f69e-43f4-a7db-d624537852d3\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.846594 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-config\") pod \"58730925-f69e-43f4-a7db-d624537852d3\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.846653 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-swift-storage-0\") pod \"58730925-f69e-43f4-a7db-d624537852d3\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.846686 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6gpt\" (UniqueName: \"kubernetes.io/projected/58730925-f69e-43f4-a7db-d624537852d3-kube-api-access-d6gpt\") pod \"58730925-f69e-43f4-a7db-d624537852d3\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.846707 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-svc\") pod \"58730925-f69e-43f4-a7db-d624537852d3\" (UID: \"58730925-f69e-43f4-a7db-d624537852d3\") " Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.864370 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58730925-f69e-43f4-a7db-d624537852d3-kube-api-access-d6gpt" (OuterVolumeSpecName: "kube-api-access-d6gpt") pod "58730925-f69e-43f4-a7db-d624537852d3" (UID: "58730925-f69e-43f4-a7db-d624537852d3"). InnerVolumeSpecName "kube-api-access-d6gpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.881525 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-config" (OuterVolumeSpecName: "config") pod "58730925-f69e-43f4-a7db-d624537852d3" (UID: "58730925-f69e-43f4-a7db-d624537852d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.884606 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "58730925-f69e-43f4-a7db-d624537852d3" (UID: "58730925-f69e-43f4-a7db-d624537852d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.889732 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "58730925-f69e-43f4-a7db-d624537852d3" (UID: "58730925-f69e-43f4-a7db-d624537852d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.890594 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "58730925-f69e-43f4-a7db-d624537852d3" (UID: "58730925-f69e-43f4-a7db-d624537852d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.905177 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "58730925-f69e-43f4-a7db-d624537852d3" (UID: "58730925-f69e-43f4-a7db-d624537852d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.948431 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.948463 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.948474 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.948486 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.948496 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6gpt\" (UniqueName: \"kubernetes.io/projected/58730925-f69e-43f4-a7db-d624537852d3-kube-api-access-d6gpt\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:43 crc kubenswrapper[4876]: I1205 08:35:43.948506 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58730925-f69e-43f4-a7db-d624537852d3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.450231 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" event={"ID":"fd5f73d3-2859-4e4d-a5d8-02acd110f11a","Type":"ContainerStarted","Data":"a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b"} Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.450671 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.455788 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549","Type":"ContainerStarted","Data":"4ef92ad678a92c40bcd4e99a31548d01e4ca48103aac27c83e04aa3a06347dcc"} Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.458725 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2de56934-652e-4571-93c2-2a504945cd26","Type":"ContainerStarted","Data":"2e8ab120ac64181d460ffa7bf6073de151c3aa4dd5d517621d9d8cfb5738b797"} Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.462644 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.463798 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-r4cl4" event={"ID":"58730925-f69e-43f4-a7db-d624537852d3","Type":"ContainerDied","Data":"a4254c2631628786986355b4a6c8f4e63e6e184ad96bb2380d852d86e162865d"} Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.463830 4876 scope.go:117] "RemoveContainer" containerID="95ae93c8b0a297929c92e8affcc66a29369ad235557b580d300b8b970be76643" Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.474997 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" podStartSLOduration=4.474974358 podStartE2EDuration="4.474974358s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:44.467324452 +0000 UTC m=+1208.955989084" watchObservedRunningTime="2025-12-05 08:35:44.474974358 +0000 UTC m=+1208.963638980" Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.536195 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-r4cl4"] Dec 05 08:35:44 crc kubenswrapper[4876]: I1205 08:35:44.550971 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-r4cl4"] Dec 05 08:35:45 crc kubenswrapper[4876]: I1205 08:35:45.477557 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549","Type":"ContainerStarted","Data":"961386face87708a1432ec2e4f8dd3ef6872b00a959a0089cffca0243201ea41"} Dec 05 08:35:45 crc kubenswrapper[4876]: I1205 08:35:45.868542 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58730925-f69e-43f4-a7db-d624537852d3" path="/var/lib/kubelet/pods/58730925-f69e-43f4-a7db-d624537852d3/volumes" Dec 05 08:35:46 crc kubenswrapper[4876]: I1205 08:35:46.493407 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2de56934-652e-4571-93c2-2a504945cd26","Type":"ContainerStarted","Data":"df093867d2ac0896237bca2e3055275c59422551fa6bc05ac47bd34cbda940f3"} Dec 05 08:35:46 crc kubenswrapper[4876]: I1205 08:35:46.493500 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerName="glance-log" containerID="cri-o://4ef92ad678a92c40bcd4e99a31548d01e4ca48103aac27c83e04aa3a06347dcc" gracePeriod=30 Dec 05 08:35:46 crc kubenswrapper[4876]: I1205 08:35:46.493659 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerName="glance-httpd" containerID="cri-o://961386face87708a1432ec2e4f8dd3ef6872b00a959a0089cffca0243201ea41" gracePeriod=30 Dec 05 08:35:46 crc kubenswrapper[4876]: I1205 08:35:46.493856 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2de56934-652e-4571-93c2-2a504945cd26" containerName="glance-log" containerID="cri-o://2e8ab120ac64181d460ffa7bf6073de151c3aa4dd5d517621d9d8cfb5738b797" gracePeriod=30 Dec 05 08:35:46 crc kubenswrapper[4876]: I1205 08:35:46.494116 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2de56934-652e-4571-93c2-2a504945cd26" containerName="glance-httpd" containerID="cri-o://df093867d2ac0896237bca2e3055275c59422551fa6bc05ac47bd34cbda940f3" gracePeriod=30 Dec 05 08:35:46 crc kubenswrapper[4876]: I1205 08:35:46.518965 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.518942023 podStartE2EDuration="6.518942023s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:46.510628648 +0000 UTC m=+1210.999293270" watchObservedRunningTime="2025-12-05 08:35:46.518942023 +0000 UTC m=+1211.007606645" Dec 05 08:35:46 crc kubenswrapper[4876]: I1205 08:35:46.543885 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.543867677 podStartE2EDuration="6.543867677s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:35:46.538774303 +0000 UTC m=+1211.027438925" watchObservedRunningTime="2025-12-05 08:35:46.543867677 +0000 UTC m=+1211.032532299" Dec 05 08:35:47 crc kubenswrapper[4876]: I1205 08:35:47.508744 4876 generic.go:334] "Generic (PLEG): container finished" podID="2de56934-652e-4571-93c2-2a504945cd26" containerID="df093867d2ac0896237bca2e3055275c59422551fa6bc05ac47bd34cbda940f3" exitCode=0 Dec 05 08:35:47 crc kubenswrapper[4876]: I1205 08:35:47.508779 4876 generic.go:334] "Generic (PLEG): container finished" podID="2de56934-652e-4571-93c2-2a504945cd26" containerID="2e8ab120ac64181d460ffa7bf6073de151c3aa4dd5d517621d9d8cfb5738b797" exitCode=143 Dec 05 08:35:47 crc kubenswrapper[4876]: I1205 08:35:47.508835 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2de56934-652e-4571-93c2-2a504945cd26","Type":"ContainerDied","Data":"df093867d2ac0896237bca2e3055275c59422551fa6bc05ac47bd34cbda940f3"} Dec 05 08:35:47 crc kubenswrapper[4876]: I1205 08:35:47.508907 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2de56934-652e-4571-93c2-2a504945cd26","Type":"ContainerDied","Data":"2e8ab120ac64181d460ffa7bf6073de151c3aa4dd5d517621d9d8cfb5738b797"} Dec 05 08:35:47 crc kubenswrapper[4876]: I1205 08:35:47.511440 4876 generic.go:334] "Generic (PLEG): container finished" podID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerID="961386face87708a1432ec2e4f8dd3ef6872b00a959a0089cffca0243201ea41" exitCode=0 Dec 05 08:35:47 crc kubenswrapper[4876]: I1205 08:35:47.511460 4876 generic.go:334] "Generic (PLEG): container finished" podID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerID="4ef92ad678a92c40bcd4e99a31548d01e4ca48103aac27c83e04aa3a06347dcc" exitCode=143 Dec 05 08:35:47 crc kubenswrapper[4876]: I1205 08:35:47.511480 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549","Type":"ContainerDied","Data":"961386face87708a1432ec2e4f8dd3ef6872b00a959a0089cffca0243201ea41"} Dec 05 08:35:47 crc kubenswrapper[4876]: I1205 08:35:47.511505 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549","Type":"ContainerDied","Data":"4ef92ad678a92c40bcd4e99a31548d01e4ca48103aac27c83e04aa3a06347dcc"} Dec 05 08:35:49 crc kubenswrapper[4876]: I1205 08:35:49.530859 4876 generic.go:334] "Generic (PLEG): container finished" podID="4c78505d-915e-4312-afff-a878017bff60" containerID="bfa61edc4a079881d538371e415544a2e93a9c3a589d05c7f54e6c19f38443ae" exitCode=0 Dec 05 08:35:49 crc kubenswrapper[4876]: I1205 08:35:49.531406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ph7r9" event={"ID":"4c78505d-915e-4312-afff-a878017bff60","Type":"ContainerDied","Data":"bfa61edc4a079881d538371e415544a2e93a9c3a589d05c7f54e6c19f38443ae"} Dec 05 08:35:51 crc kubenswrapper[4876]: I1205 08:35:51.484843 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:35:51 crc kubenswrapper[4876]: I1205 08:35:51.545239 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hchbt"] Dec 05 08:35:51 crc kubenswrapper[4876]: I1205 08:35:51.545536 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="dnsmasq-dns" containerID="cri-o://08e18a87af24b5174714da524d5711889d67739069c67b03d6397d683361086b" gracePeriod=10 Dec 05 08:35:52 crc kubenswrapper[4876]: I1205 08:35:52.104183 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: connect: connection refused" Dec 05 08:35:52 crc kubenswrapper[4876]: I1205 08:35:52.571186 4876 generic.go:334] "Generic (PLEG): container finished" podID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerID="08e18a87af24b5174714da524d5711889d67739069c67b03d6397d683361086b" exitCode=0 Dec 05 08:35:52 crc kubenswrapper[4876]: I1205 08:35:52.571236 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" event={"ID":"6e7c039f-5af6-4394-b41b-9d8273912ddb","Type":"ContainerDied","Data":"08e18a87af24b5174714da524d5711889d67739069c67b03d6397d683361086b"} Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.045267 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.226701 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-logs\") pod \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.226792 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.226853 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-combined-ca-bundle\") pod \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.226920 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdbcw\" (UniqueName: \"kubernetes.io/projected/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-kube-api-access-hdbcw\") pod \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.226995 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-scripts\") pod \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.227017 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-httpd-run\") pod \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.227057 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-public-tls-certs\") pod \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.227508 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" (UID: "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.227547 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-logs" (OuterVolumeSpecName: "logs") pod "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" (UID: "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.228089 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-config-data\") pod \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\" (UID: \"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549\") " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.228862 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.228880 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.233451 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" (UID: "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.233705 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-kube-api-access-hdbcw" (OuterVolumeSpecName: "kube-api-access-hdbcw") pod "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" (UID: "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549"). InnerVolumeSpecName "kube-api-access-hdbcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.235330 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-scripts" (OuterVolumeSpecName: "scripts") pod "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" (UID: "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.263182 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" (UID: "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.279135 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" (UID: "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.283105 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-config-data" (OuterVolumeSpecName: "config-data") pod "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" (UID: "fbe85d9f-b9b0-47dc-b53b-80c4eff5b549"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.329873 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.329957 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.329975 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.329989 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdbcw\" (UniqueName: \"kubernetes.io/projected/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-kube-api-access-hdbcw\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.330000 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.330012 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.356425 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.431271 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.579884 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fbe85d9f-b9b0-47dc-b53b-80c4eff5b549","Type":"ContainerDied","Data":"f37b0b4e77f6abba7ef0ebbdddf944b021fab7324eca36ff9546ad1892b68215"} Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.580169 4876 scope.go:117] "RemoveContainer" containerID="961386face87708a1432ec2e4f8dd3ef6872b00a959a0089cffca0243201ea41" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.580003 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.634095 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.646512 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.679293 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:35:53 crc kubenswrapper[4876]: E1205 08:35:53.679673 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerName="glance-log" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.679692 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerName="glance-log" Dec 05 08:35:53 crc kubenswrapper[4876]: E1205 08:35:53.679723 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58730925-f69e-43f4-a7db-d624537852d3" containerName="init" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.679729 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="58730925-f69e-43f4-a7db-d624537852d3" containerName="init" Dec 05 08:35:53 crc kubenswrapper[4876]: E1205 08:35:53.679744 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerName="glance-httpd" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.679750 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerName="glance-httpd" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.679927 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerName="glance-log" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.679947 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="58730925-f69e-43f4-a7db-d624537852d3" containerName="init" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.679956 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" containerName="glance-httpd" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.680790 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.683199 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.687475 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.694014 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.835804 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbe85d9f-b9b0-47dc-b53b-80c4eff5b549" path="/var/lib/kubelet/pods/fbe85d9f-b9b0-47dc-b53b-80c4eff5b549/volumes" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.844381 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.844453 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xphnv\" (UniqueName: \"kubernetes.io/projected/c328d736-74fe-4075-b291-ea6ff38ef1ef-kube-api-access-xphnv\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.844480 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.844697 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-config-data\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.844739 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-logs\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.844839 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.844866 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-scripts\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.844943 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.955766 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.955843 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-scripts\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.955946 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.956056 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.956097 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xphnv\" (UniqueName: \"kubernetes.io/projected/c328d736-74fe-4075-b291-ea6ff38ef1ef-kube-api-access-xphnv\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.956117 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.956145 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-config-data\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.956194 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-logs\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.957192 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-logs\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.957569 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.959631 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.963289 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.963444 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-scripts\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.966945 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.967476 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-config-data\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.977627 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xphnv\" (UniqueName: \"kubernetes.io/projected/c328d736-74fe-4075-b291-ea6ff38ef1ef-kube-api-access-xphnv\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:53 crc kubenswrapper[4876]: I1205 08:35:53.997684 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " pod="openstack/glance-default-external-api-0" Dec 05 08:35:54 crc kubenswrapper[4876]: I1205 08:35:54.300608 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:35:59 crc kubenswrapper[4876]: E1205 08:35:59.168051 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 05 08:35:59 crc kubenswrapper[4876]: E1205 08:35:59.168724 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5cbh77h64fh544hb4h677h55dh677h54fh676h5c6h7bh5d8hbch68h548h79h7bh585hf5h5b6h6ch659h54bh678h64dh648h57ch584h5f6h7h88q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m5qgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(ee3139f2-0c55-464e-bdbf-eefcc4bb59c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.251189 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.269276 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-scripts\") pod \"4c78505d-915e-4312-afff-a878017bff60\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.269464 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-config-data\") pod \"4c78505d-915e-4312-afff-a878017bff60\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.269578 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwbjh\" (UniqueName: \"kubernetes.io/projected/4c78505d-915e-4312-afff-a878017bff60-kube-api-access-zwbjh\") pod \"4c78505d-915e-4312-afff-a878017bff60\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.269607 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-combined-ca-bundle\") pod \"4c78505d-915e-4312-afff-a878017bff60\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.269629 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-fernet-keys\") pod \"4c78505d-915e-4312-afff-a878017bff60\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.269700 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-credential-keys\") pod \"4c78505d-915e-4312-afff-a878017bff60\" (UID: \"4c78505d-915e-4312-afff-a878017bff60\") " Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.275999 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-scripts" (OuterVolumeSpecName: "scripts") pod "4c78505d-915e-4312-afff-a878017bff60" (UID: "4c78505d-915e-4312-afff-a878017bff60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.276703 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c78505d-915e-4312-afff-a878017bff60-kube-api-access-zwbjh" (OuterVolumeSpecName: "kube-api-access-zwbjh") pod "4c78505d-915e-4312-afff-a878017bff60" (UID: "4c78505d-915e-4312-afff-a878017bff60"). InnerVolumeSpecName "kube-api-access-zwbjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.277284 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4c78505d-915e-4312-afff-a878017bff60" (UID: "4c78505d-915e-4312-afff-a878017bff60"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.277780 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4c78505d-915e-4312-afff-a878017bff60" (UID: "4c78505d-915e-4312-afff-a878017bff60"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.304962 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c78505d-915e-4312-afff-a878017bff60" (UID: "4c78505d-915e-4312-afff-a878017bff60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.309412 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-config-data" (OuterVolumeSpecName: "config-data") pod "4c78505d-915e-4312-afff-a878017bff60" (UID: "4c78505d-915e-4312-afff-a878017bff60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.371269 4876 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.371309 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.371321 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.371336 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwbjh\" (UniqueName: \"kubernetes.io/projected/4c78505d-915e-4312-afff-a878017bff60-kube-api-access-zwbjh\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.371350 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.371359 4876 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c78505d-915e-4312-afff-a878017bff60-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.642722 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ph7r9" event={"ID":"4c78505d-915e-4312-afff-a878017bff60","Type":"ContainerDied","Data":"5f4312d0d5be88d9991a3c3b9971e25ce438331bc3387b2a3435238a92839fc2"} Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.642768 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f4312d0d5be88d9991a3c3b9971e25ce438331bc3387b2a3435238a92839fc2" Dec 05 08:35:59 crc kubenswrapper[4876]: I1205 08:35:59.642821 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ph7r9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.328334 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ph7r9"] Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.337152 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ph7r9"] Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.445138 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wc9k9"] Dec 05 08:36:00 crc kubenswrapper[4876]: E1205 08:36:00.445559 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c78505d-915e-4312-afff-a878017bff60" containerName="keystone-bootstrap" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.445577 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c78505d-915e-4312-afff-a878017bff60" containerName="keystone-bootstrap" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.445787 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c78505d-915e-4312-afff-a878017bff60" containerName="keystone-bootstrap" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.446424 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.450256 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.450493 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.450547 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.450506 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.450730 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fmb5b" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.462416 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wc9k9"] Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.491651 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-config-data\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.491714 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-combined-ca-bundle\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.491755 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwlrc\" (UniqueName: \"kubernetes.io/projected/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-kube-api-access-hwlrc\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.491791 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-credential-keys\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.491818 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-fernet-keys\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.491841 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-scripts\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.593410 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwlrc\" (UniqueName: \"kubernetes.io/projected/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-kube-api-access-hwlrc\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.593465 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-credential-keys\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.593491 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-fernet-keys\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.593516 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-scripts\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.593604 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-config-data\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.593636 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-combined-ca-bundle\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.598091 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-config-data\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.598475 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-credential-keys\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.610688 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwlrc\" (UniqueName: \"kubernetes.io/projected/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-kube-api-access-hwlrc\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.618792 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-scripts\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.619753 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-fernet-keys\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.620352 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-combined-ca-bundle\") pod \"keystone-bootstrap-wc9k9\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:00 crc kubenswrapper[4876]: I1205 08:36:00.779309 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:01 crc kubenswrapper[4876]: I1205 08:36:01.847599 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c78505d-915e-4312-afff-a878017bff60" path="/var/lib/kubelet/pods/4c78505d-915e-4312-afff-a878017bff60/volumes" Dec 05 08:36:02 crc kubenswrapper[4876]: I1205 08:36:02.104199 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: i/o timeout" Dec 05 08:36:06 crc kubenswrapper[4876]: E1205 08:36:06.464436 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 05 08:36:06 crc kubenswrapper[4876]: E1205 08:36:06.465130 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2nptt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-xbjb2_openstack(5a1ae752-b331-4d0f-8039-e5720312f674): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:36:06 crc kubenswrapper[4876]: E1205 08:36:06.467070 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-xbjb2" podUID="5a1ae752-b331-4d0f-8039-e5720312f674" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.514106 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.598331 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-sb\") pod \"6e7c039f-5af6-4394-b41b-9d8273912ddb\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.598369 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxf5f\" (UniqueName: \"kubernetes.io/projected/6e7c039f-5af6-4394-b41b-9d8273912ddb-kube-api-access-fxf5f\") pod \"6e7c039f-5af6-4394-b41b-9d8273912ddb\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.598399 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-nb\") pod \"6e7c039f-5af6-4394-b41b-9d8273912ddb\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.598424 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-config\") pod \"6e7c039f-5af6-4394-b41b-9d8273912ddb\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.598441 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-swift-storage-0\") pod \"6e7c039f-5af6-4394-b41b-9d8273912ddb\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.598459 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-svc\") pod \"6e7c039f-5af6-4394-b41b-9d8273912ddb\" (UID: \"6e7c039f-5af6-4394-b41b-9d8273912ddb\") " Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.604195 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e7c039f-5af6-4394-b41b-9d8273912ddb-kube-api-access-fxf5f" (OuterVolumeSpecName: "kube-api-access-fxf5f") pod "6e7c039f-5af6-4394-b41b-9d8273912ddb" (UID: "6e7c039f-5af6-4394-b41b-9d8273912ddb"). InnerVolumeSpecName "kube-api-access-fxf5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.637852 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6e7c039f-5af6-4394-b41b-9d8273912ddb" (UID: "6e7c039f-5af6-4394-b41b-9d8273912ddb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.638633 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6e7c039f-5af6-4394-b41b-9d8273912ddb" (UID: "6e7c039f-5af6-4394-b41b-9d8273912ddb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.642413 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-config" (OuterVolumeSpecName: "config") pod "6e7c039f-5af6-4394-b41b-9d8273912ddb" (UID: "6e7c039f-5af6-4394-b41b-9d8273912ddb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.642887 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6e7c039f-5af6-4394-b41b-9d8273912ddb" (UID: "6e7c039f-5af6-4394-b41b-9d8273912ddb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.651417 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6e7c039f-5af6-4394-b41b-9d8273912ddb" (UID: "6e7c039f-5af6-4394-b41b-9d8273912ddb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.699884 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.699973 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxf5f\" (UniqueName: \"kubernetes.io/projected/6e7c039f-5af6-4394-b41b-9d8273912ddb-kube-api-access-fxf5f\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.699986 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.699997 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.700010 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.700022 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e7c039f-5af6-4394-b41b-9d8273912ddb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.701561 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" event={"ID":"6e7c039f-5af6-4394-b41b-9d8273912ddb","Type":"ContainerDied","Data":"b1f0b3239dec7e77ec5389ad5563909a5ef6e206910bf7744efba04c0232a8bd"} Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.701617 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.704005 4876 generic.go:334] "Generic (PLEG): container finished" podID="207c5e9c-418a-4c07-8e8a-bb79948d855d" containerID="cbc720abbc55b77c1cbdf9bb522e82ac772b3b46fdfb8f291d1fb783bcaf5ad2" exitCode=0 Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.704119 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2rwzp" event={"ID":"207c5e9c-418a-4c07-8e8a-bb79948d855d","Type":"ContainerDied","Data":"cbc720abbc55b77c1cbdf9bb522e82ac772b3b46fdfb8f291d1fb783bcaf5ad2"} Dec 05 08:36:06 crc kubenswrapper[4876]: E1205 08:36:06.705248 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-xbjb2" podUID="5a1ae752-b331-4d0f-8039-e5720312f674" Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.764769 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hchbt"] Dec 05 08:36:06 crc kubenswrapper[4876]: I1205 08:36:06.774077 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hchbt"] Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.105034 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-hchbt" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: i/o timeout" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.578011 4876 scope.go:117] "RemoveContainer" containerID="4ef92ad678a92c40bcd4e99a31548d01e4ca48103aac27c83e04aa3a06347dcc" Dec 05 08:36:07 crc kubenswrapper[4876]: E1205 08:36:07.636646 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 05 08:36:07 crc kubenswrapper[4876]: E1205 08:36:07.636845 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55btw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vxcln_openstack(62f1d7bc-1a36-4fc0-a94d-72fcedda052e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 08:36:07 crc kubenswrapper[4876]: E1205 08:36:07.638174 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vxcln" podUID="62f1d7bc-1a36-4fc0-a94d-72fcedda052e" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.669378 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.722266 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2de56934-652e-4571-93c2-2a504945cd26","Type":"ContainerDied","Data":"1fed6a4d40221ef202049ce908d3be11c86733e601a414dca3fe77b7df788474"} Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.722292 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:07 crc kubenswrapper[4876]: E1205 08:36:07.733022 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-vxcln" podUID="62f1d7bc-1a36-4fc0-a94d-72fcedda052e" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.816921 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-internal-tls-certs\") pod \"2de56934-652e-4571-93c2-2a504945cd26\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.817323 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-httpd-run\") pod \"2de56934-652e-4571-93c2-2a504945cd26\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.817598 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-combined-ca-bundle\") pod \"2de56934-652e-4571-93c2-2a504945cd26\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.818185 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-config-data\") pod \"2de56934-652e-4571-93c2-2a504945cd26\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.817822 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2de56934-652e-4571-93c2-2a504945cd26" (UID: "2de56934-652e-4571-93c2-2a504945cd26"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.818363 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-logs\") pod \"2de56934-652e-4571-93c2-2a504945cd26\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.818446 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"2de56934-652e-4571-93c2-2a504945cd26\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.818486 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-scripts\") pod \"2de56934-652e-4571-93c2-2a504945cd26\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.818527 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k8sb\" (UniqueName: \"kubernetes.io/projected/2de56934-652e-4571-93c2-2a504945cd26-kube-api-access-9k8sb\") pod \"2de56934-652e-4571-93c2-2a504945cd26\" (UID: \"2de56934-652e-4571-93c2-2a504945cd26\") " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.819213 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-logs" (OuterVolumeSpecName: "logs") pod "2de56934-652e-4571-93c2-2a504945cd26" (UID: "2de56934-652e-4571-93c2-2a504945cd26"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.819801 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.819836 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2de56934-652e-4571-93c2-2a504945cd26-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.822832 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-scripts" (OuterVolumeSpecName: "scripts") pod "2de56934-652e-4571-93c2-2a504945cd26" (UID: "2de56934-652e-4571-93c2-2a504945cd26"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.822940 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de56934-652e-4571-93c2-2a504945cd26-kube-api-access-9k8sb" (OuterVolumeSpecName: "kube-api-access-9k8sb") pod "2de56934-652e-4571-93c2-2a504945cd26" (UID: "2de56934-652e-4571-93c2-2a504945cd26"). InnerVolumeSpecName "kube-api-access-9k8sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.836048 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" path="/var/lib/kubelet/pods/6e7c039f-5af6-4394-b41b-9d8273912ddb/volumes" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.836260 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "2de56934-652e-4571-93c2-2a504945cd26" (UID: "2de56934-652e-4571-93c2-2a504945cd26"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.863204 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2de56934-652e-4571-93c2-2a504945cd26" (UID: "2de56934-652e-4571-93c2-2a504945cd26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.874989 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-config-data" (OuterVolumeSpecName: "config-data") pod "2de56934-652e-4571-93c2-2a504945cd26" (UID: "2de56934-652e-4571-93c2-2a504945cd26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.916213 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2de56934-652e-4571-93c2-2a504945cd26" (UID: "2de56934-652e-4571-93c2-2a504945cd26"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.920738 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.920817 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.921031 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.921059 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.921067 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2de56934-652e-4571-93c2-2a504945cd26-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.921076 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k8sb\" (UniqueName: \"kubernetes.io/projected/2de56934-652e-4571-93c2-2a504945cd26-kube-api-access-9k8sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:07 crc kubenswrapper[4876]: I1205 08:36:07.937872 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.022774 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.057335 4876 scope.go:117] "RemoveContainer" containerID="08e18a87af24b5174714da524d5711889d67739069c67b03d6397d683361086b" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.058512 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.077691 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.089515 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:36:08 crc kubenswrapper[4876]: E1205 08:36:08.090026 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de56934-652e-4571-93c2-2a504945cd26" containerName="glance-log" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.090051 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de56934-652e-4571-93c2-2a504945cd26" containerName="glance-log" Dec 05 08:36:08 crc kubenswrapper[4876]: E1205 08:36:08.090086 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="init" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.090095 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="init" Dec 05 08:36:08 crc kubenswrapper[4876]: E1205 08:36:08.090111 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de56934-652e-4571-93c2-2a504945cd26" containerName="glance-httpd" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.090118 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de56934-652e-4571-93c2-2a504945cd26" containerName="glance-httpd" Dec 05 08:36:08 crc kubenswrapper[4876]: E1205 08:36:08.090131 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="dnsmasq-dns" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.090138 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="dnsmasq-dns" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.090495 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de56934-652e-4571-93c2-2a504945cd26" containerName="glance-log" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.090521 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de56934-652e-4571-93c2-2a504945cd26" containerName="glance-httpd" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.090544 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e7c039f-5af6-4394-b41b-9d8273912ddb" containerName="dnsmasq-dns" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.091636 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.094067 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.094233 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.099747 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.147816 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.174507 4876 scope.go:117] "RemoveContainer" containerID="c6703baa6ac24ed7638811d50e540df6ad4972e57ef9a9723c2ff56664b78e38" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.228873 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.229182 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.229234 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-logs\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.229263 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.229305 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.229325 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.229352 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m97fp\" (UniqueName: \"kubernetes.io/projected/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-kube-api-access-m97fp\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.229375 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.258742 4876 scope.go:117] "RemoveContainer" containerID="df093867d2ac0896237bca2e3055275c59422551fa6bc05ac47bd34cbda940f3" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.292137 4876 scope.go:117] "RemoveContainer" containerID="2e8ab120ac64181d460ffa7bf6073de151c3aa4dd5d517621d9d8cfb5738b797" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330187 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-combined-ca-bundle\") pod \"207c5e9c-418a-4c07-8e8a-bb79948d855d\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330225 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs9gh\" (UniqueName: \"kubernetes.io/projected/207c5e9c-418a-4c07-8e8a-bb79948d855d-kube-api-access-gs9gh\") pod \"207c5e9c-418a-4c07-8e8a-bb79948d855d\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330268 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-config\") pod \"207c5e9c-418a-4c07-8e8a-bb79948d855d\" (UID: \"207c5e9c-418a-4c07-8e8a-bb79948d855d\") " Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330607 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330646 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m97fp\" (UniqueName: \"kubernetes.io/projected/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-kube-api-access-m97fp\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330673 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330715 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330773 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330816 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-logs\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330849 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.330949 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.331277 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.332408 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-logs\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.332725 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.337005 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.338035 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.340801 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.341113 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.341994 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207c5e9c-418a-4c07-8e8a-bb79948d855d-kube-api-access-gs9gh" (OuterVolumeSpecName: "kube-api-access-gs9gh") pod "207c5e9c-418a-4c07-8e8a-bb79948d855d" (UID: "207c5e9c-418a-4c07-8e8a-bb79948d855d"). InnerVolumeSpecName "kube-api-access-gs9gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.347222 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m97fp\" (UniqueName: \"kubernetes.io/projected/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-kube-api-access-m97fp\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.380623 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.414112 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.418261 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-config" (OuterVolumeSpecName: "config") pod "207c5e9c-418a-4c07-8e8a-bb79948d855d" (UID: "207c5e9c-418a-4c07-8e8a-bb79948d855d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.423650 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "207c5e9c-418a-4c07-8e8a-bb79948d855d" (UID: "207c5e9c-418a-4c07-8e8a-bb79948d855d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.432124 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.432279 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs9gh\" (UniqueName: \"kubernetes.io/projected/207c5e9c-418a-4c07-8e8a-bb79948d855d-kube-api-access-gs9gh\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.432352 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/207c5e9c-418a-4c07-8e8a-bb79948d855d-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.575549 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wc9k9"] Dec 05 08:36:08 crc kubenswrapper[4876]: W1205 08:36:08.589833 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9187c3bd_2fa8_41dd_a669_09d0a4fd7cd1.slice/crio-a95c9a423b951740ee902f65e7daa34a20c0d4e65944352bc5311b6a96cb47ef WatchSource:0}: Error finding container a95c9a423b951740ee902f65e7daa34a20c0d4e65944352bc5311b6a96cb47ef: Status 404 returned error can't find the container with id a95c9a423b951740ee902f65e7daa34a20c0d4e65944352bc5311b6a96cb47ef Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.623083 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.751699 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2rwzp" event={"ID":"207c5e9c-418a-4c07-8e8a-bb79948d855d","Type":"ContainerDied","Data":"1e0aae61edfe35977d5420788905b8ae21c239343e76bc171506193a8bc75c9a"} Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.752095 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0aae61edfe35977d5420788905b8ae21c239343e76bc171506193a8bc75c9a" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.754461 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2rwzp" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.766279 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n5bmx" event={"ID":"460f93bd-e732-42c0-8e75-625f3346800c","Type":"ContainerStarted","Data":"d5b7822ecdb59c6b8b44f0f4c15198949fd589345f3891c5cb7760da54614320"} Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.770781 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-m2khz" event={"ID":"2d18b829-e516-41de-989d-bea32b940409","Type":"ContainerStarted","Data":"ba7f7adc5c72b82dec3c9ece793de8e91289022fd9785b8cf8677e04284a946e"} Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.771752 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c328d736-74fe-4075-b291-ea6ff38ef1ef","Type":"ContainerStarted","Data":"94f9a599f553ae6e0eab06a7f024790e0660e107db9042267aeb67e6f4fb61ac"} Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.774250 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5","Type":"ContainerStarted","Data":"bdb2749c0bdbf1d92e5bea8bf5cb1424c07fab5d7e2895e4bfcb0f9ab6f0c4d8"} Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.778358 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wc9k9" event={"ID":"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1","Type":"ContainerStarted","Data":"a95c9a423b951740ee902f65e7daa34a20c0d4e65944352bc5311b6a96cb47ef"} Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.786659 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-n5bmx" podStartSLOduration=3.218415951 podStartE2EDuration="28.785917163s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="2025-12-05 08:35:42.4758784 +0000 UTC m=+1206.964543022" lastFinishedPulling="2025-12-05 08:36:08.043379612 +0000 UTC m=+1232.532044234" observedRunningTime="2025-12-05 08:36:08.782604808 +0000 UTC m=+1233.271269440" watchObservedRunningTime="2025-12-05 08:36:08.785917163 +0000 UTC m=+1233.274581785" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.823657 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-m2khz" podStartSLOduration=3.105896381 podStartE2EDuration="28.823638579s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="2025-12-05 08:35:42.32830805 +0000 UTC m=+1206.816972672" lastFinishedPulling="2025-12-05 08:36:08.046050248 +0000 UTC m=+1232.534714870" observedRunningTime="2025-12-05 08:36:08.821973021 +0000 UTC m=+1233.310637673" watchObservedRunningTime="2025-12-05 08:36:08.823638579 +0000 UTC m=+1233.312303201" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.878069 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-4qckc"] Dec 05 08:36:08 crc kubenswrapper[4876]: E1205 08:36:08.878389 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207c5e9c-418a-4c07-8e8a-bb79948d855d" containerName="neutron-db-sync" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.878402 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="207c5e9c-418a-4c07-8e8a-bb79948d855d" containerName="neutron-db-sync" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.878573 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="207c5e9c-418a-4c07-8e8a-bb79948d855d" containerName="neutron-db-sync" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.893515 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.899471 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-4qckc"] Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.983617 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-75d658964-r6f2m"] Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.985106 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.989314 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cd7mp" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.989487 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.989828 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.989963 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 08:36:08 crc kubenswrapper[4876]: I1205 08:36:08.997622 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75d658964-r6f2m"] Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.034397 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:36:09 crc kubenswrapper[4876]: W1205 08:36:09.044507 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69166e3f_0a62_4a36_9fd4_2a78e35c43b6.slice/crio-3c1e3f37b7198dd537189638ecde00a7c086558583161c75e8dd05fba52f23fc WatchSource:0}: Error finding container 3c1e3f37b7198dd537189638ecde00a7c086558583161c75e8dd05fba52f23fc: Status 404 returned error can't find the container with id 3c1e3f37b7198dd537189638ecde00a7c086558583161c75e8dd05fba52f23fc Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.048228 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.048306 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.048352 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rctk6\" (UniqueName: \"kubernetes.io/projected/7eb4b169-060a-45b0-8607-7538547b5640-kube-api-access-rctk6\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.048385 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.048691 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.048733 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-config\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.150736 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.150841 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.150935 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rctk6\" (UniqueName: \"kubernetes.io/projected/7eb4b169-060a-45b0-8607-7538547b5640-kube-api-access-rctk6\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.150974 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151036 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9chf\" (UniqueName: \"kubernetes.io/projected/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-kube-api-access-c9chf\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151085 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151113 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-combined-ca-bundle\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151134 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-config\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151171 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-ovndb-tls-certs\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151193 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-httpd-config\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151258 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-config\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151644 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.151936 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.152650 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.153053 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-config\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.153186 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.189977 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rctk6\" (UniqueName: \"kubernetes.io/projected/7eb4b169-060a-45b0-8607-7538547b5640-kube-api-access-rctk6\") pod \"dnsmasq-dns-5ccc5c4795-4qckc\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.252551 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.255166 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-config\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.261251 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-config\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.262171 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9chf\" (UniqueName: \"kubernetes.io/projected/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-kube-api-access-c9chf\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.262232 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-combined-ca-bundle\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.262268 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-ovndb-tls-certs\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.262290 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-httpd-config\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.275104 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-combined-ca-bundle\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.304059 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-ovndb-tls-certs\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.304815 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-httpd-config\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.321776 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9chf\" (UniqueName: \"kubernetes.io/projected/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-kube-api-access-c9chf\") pod \"neutron-75d658964-r6f2m\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.331081 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.792308 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-4qckc"] Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.798937 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c328d736-74fe-4075-b291-ea6ff38ef1ef","Type":"ContainerStarted","Data":"967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3"} Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.800088 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69166e3f-0a62-4a36-9fd4-2a78e35c43b6","Type":"ContainerStarted","Data":"3c1e3f37b7198dd537189638ecde00a7c086558583161c75e8dd05fba52f23fc"} Dec 05 08:36:09 crc kubenswrapper[4876]: W1205 08:36:09.801293 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eb4b169_060a_45b0_8607_7538547b5640.slice/crio-08bc39ed30a8c882534c885d74a0013415387dc83b5acd6b5e7209727cdd6e79 WatchSource:0}: Error finding container 08bc39ed30a8c882534c885d74a0013415387dc83b5acd6b5e7209727cdd6e79: Status 404 returned error can't find the container with id 08bc39ed30a8c882534c885d74a0013415387dc83b5acd6b5e7209727cdd6e79 Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.805401 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wc9k9" event={"ID":"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1","Type":"ContainerStarted","Data":"d0694e5b242890e48a9b0af47bb44693b50de3adca8d565e9f7a01d0e76ac914"} Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.842356 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wc9k9" podStartSLOduration=9.842335437 podStartE2EDuration="9.842335437s" podCreationTimestamp="2025-12-05 08:36:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:09.830300184 +0000 UTC m=+1234.318964806" watchObservedRunningTime="2025-12-05 08:36:09.842335437 +0000 UTC m=+1234.331000059" Dec 05 08:36:09 crc kubenswrapper[4876]: I1205 08:36:09.845977 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2de56934-652e-4571-93c2-2a504945cd26" path="/var/lib/kubelet/pods/2de56934-652e-4571-93c2-2a504945cd26/volumes" Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.111636 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75d658964-r6f2m"] Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.888799 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75d658964-r6f2m" event={"ID":"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0","Type":"ContainerStarted","Data":"6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc"} Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.889252 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75d658964-r6f2m" event={"ID":"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0","Type":"ContainerStarted","Data":"65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362"} Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.889270 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75d658964-r6f2m" event={"ID":"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0","Type":"ContainerStarted","Data":"64fc9f8a6aae6fb1d89781e4d3af3b455d6fb467d5edb3ce5fae1cae2011f061"} Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.889311 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.898261 4876 generic.go:334] "Generic (PLEG): container finished" podID="7eb4b169-060a-45b0-8607-7538547b5640" containerID="5c0257f1ae3521b305faf224679b6a31da84614003320d50a3aef367435f9c7c" exitCode=0 Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.899042 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" event={"ID":"7eb4b169-060a-45b0-8607-7538547b5640","Type":"ContainerDied","Data":"5c0257f1ae3521b305faf224679b6a31da84614003320d50a3aef367435f9c7c"} Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.899073 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" event={"ID":"7eb4b169-060a-45b0-8607-7538547b5640","Type":"ContainerStarted","Data":"08bc39ed30a8c882534c885d74a0013415387dc83b5acd6b5e7209727cdd6e79"} Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.904686 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c328d736-74fe-4075-b291-ea6ff38ef1ef","Type":"ContainerStarted","Data":"24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6"} Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.921558 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-75d658964-r6f2m" podStartSLOduration=2.921538901 podStartE2EDuration="2.921538901s" podCreationTimestamp="2025-12-05 08:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:10.911593817 +0000 UTC m=+1235.400258439" watchObservedRunningTime="2025-12-05 08:36:10.921538901 +0000 UTC m=+1235.410203523" Dec 05 08:36:10 crc kubenswrapper[4876]: I1205 08:36:10.924032 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69166e3f-0a62-4a36-9fd4-2a78e35c43b6","Type":"ContainerStarted","Data":"4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731"} Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.231718 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=18.231706728 podStartE2EDuration="18.231706728s" podCreationTimestamp="2025-12-05 08:35:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:11.231432521 +0000 UTC m=+1235.720097143" watchObservedRunningTime="2025-12-05 08:36:11.231706728 +0000 UTC m=+1235.720371350" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.686775 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5ff8f8c8c7-rx6w7"] Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.697459 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.704553 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.704932 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.714411 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5ff8f8c8c7-rx6w7"] Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.771552 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-combined-ca-bundle\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.771587 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-internal-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.771621 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-ovndb-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.771660 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlrz9\" (UniqueName: \"kubernetes.io/projected/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-kube-api-access-xlrz9\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.771692 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-httpd-config\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.771718 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-config\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.771755 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-public-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.872975 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-httpd-config\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.873050 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-config\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.873110 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-public-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.873265 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-combined-ca-bundle\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.873291 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-internal-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.873323 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-ovndb-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.873367 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlrz9\" (UniqueName: \"kubernetes.io/projected/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-kube-api-access-xlrz9\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.879414 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-httpd-config\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.882543 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-internal-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.884701 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-combined-ca-bundle\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.886801 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-config\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.887008 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-ovndb-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.894772 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-public-tls-certs\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.896570 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlrz9\" (UniqueName: \"kubernetes.io/projected/cd9bce2b-4ecb-45b0-a1b2-309c3f450a47-kube-api-access-xlrz9\") pod \"neutron-5ff8f8c8c7-rx6w7\" (UID: \"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47\") " pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.941734 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69166e3f-0a62-4a36-9fd4-2a78e35c43b6","Type":"ContainerStarted","Data":"cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053"} Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.945497 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" event={"ID":"7eb4b169-060a-45b0-8607-7538547b5640","Type":"ContainerStarted","Data":"35822f26df723a06788493800f3466d8a9149a0df877825b531946c6eb9868d6"} Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.946123 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.978602 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.978577593 podStartE2EDuration="3.978577593s" podCreationTimestamp="2025-12-05 08:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:11.965449248 +0000 UTC m=+1236.454113880" watchObservedRunningTime="2025-12-05 08:36:11.978577593 +0000 UTC m=+1236.467242215" Dec 05 08:36:11 crc kubenswrapper[4876]: I1205 08:36:11.987296 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" podStartSLOduration=3.9872735009999998 podStartE2EDuration="3.987273501s" podCreationTimestamp="2025-12-05 08:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:11.983660288 +0000 UTC m=+1236.472324930" watchObservedRunningTime="2025-12-05 08:36:11.987273501 +0000 UTC m=+1236.475938123" Dec 05 08:36:12 crc kubenswrapper[4876]: I1205 08:36:12.074335 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:12 crc kubenswrapper[4876]: I1205 08:36:12.538887 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5ff8f8c8c7-rx6w7"] Dec 05 08:36:12 crc kubenswrapper[4876]: I1205 08:36:12.965387 4876 generic.go:334] "Generic (PLEG): container finished" podID="460f93bd-e732-42c0-8e75-625f3346800c" containerID="d5b7822ecdb59c6b8b44f0f4c15198949fd589345f3891c5cb7760da54614320" exitCode=0 Dec 05 08:36:12 crc kubenswrapper[4876]: I1205 08:36:12.965966 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n5bmx" event={"ID":"460f93bd-e732-42c0-8e75-625f3346800c","Type":"ContainerDied","Data":"d5b7822ecdb59c6b8b44f0f4c15198949fd589345f3891c5cb7760da54614320"} Dec 05 08:36:12 crc kubenswrapper[4876]: I1205 08:36:12.970966 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ff8f8c8c7-rx6w7" event={"ID":"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47","Type":"ContainerStarted","Data":"f0689d1e438524a32fd79150258a26200f5d38a604db20b9d740af1f2e8eaafe"} Dec 05 08:36:12 crc kubenswrapper[4876]: I1205 08:36:12.971009 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ff8f8c8c7-rx6w7" event={"ID":"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47","Type":"ContainerStarted","Data":"b18f4c94124d92c96ebbfb8e65742631585ac5c813cc618e82d8da7dada7cb0c"} Dec 05 08:36:14 crc kubenswrapper[4876]: I1205 08:36:14.002705 4876 generic.go:334] "Generic (PLEG): container finished" podID="9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" containerID="d0694e5b242890e48a9b0af47bb44693b50de3adca8d565e9f7a01d0e76ac914" exitCode=0 Dec 05 08:36:14 crc kubenswrapper[4876]: I1205 08:36:14.003239 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wc9k9" event={"ID":"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1","Type":"ContainerDied","Data":"d0694e5b242890e48a9b0af47bb44693b50de3adca8d565e9f7a01d0e76ac914"} Dec 05 08:36:14 crc kubenswrapper[4876]: I1205 08:36:14.301034 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 08:36:14 crc kubenswrapper[4876]: I1205 08:36:14.301402 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 08:36:14 crc kubenswrapper[4876]: I1205 08:36:14.343657 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 08:36:14 crc kubenswrapper[4876]: I1205 08:36:14.359207 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 08:36:15 crc kubenswrapper[4876]: I1205 08:36:15.013602 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 08:36:15 crc kubenswrapper[4876]: I1205 08:36:15.013650 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.020557 4876 generic.go:334] "Generic (PLEG): container finished" podID="2d18b829-e516-41de-989d-bea32b940409" containerID="ba7f7adc5c72b82dec3c9ece793de8e91289022fd9785b8cf8677e04284a946e" exitCode=0 Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.020679 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-m2khz" event={"ID":"2d18b829-e516-41de-989d-bea32b940409","Type":"ContainerDied","Data":"ba7f7adc5c72b82dec3c9ece793de8e91289022fd9785b8cf8677e04284a946e"} Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.793568 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.950309 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-scripts\") pod \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.950366 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-credential-keys\") pod \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.950435 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-fernet-keys\") pod \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.950485 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwlrc\" (UniqueName: \"kubernetes.io/projected/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-kube-api-access-hwlrc\") pod \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.950541 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-combined-ca-bundle\") pod \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.950640 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-config-data\") pod \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\" (UID: \"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1\") " Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.956235 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" (UID: "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.956811 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-scripts" (OuterVolumeSpecName: "scripts") pod "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" (UID: "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.959667 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-kube-api-access-hwlrc" (OuterVolumeSpecName: "kube-api-access-hwlrc") pod "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" (UID: "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1"). InnerVolumeSpecName "kube-api-access-hwlrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.976182 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" (UID: "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.976803 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-config-data" (OuterVolumeSpecName: "config-data") pod "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" (UID: "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:16 crc kubenswrapper[4876]: I1205 08:36:16.978687 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" (UID: "9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.030499 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wc9k9" event={"ID":"9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1","Type":"ContainerDied","Data":"a95c9a423b951740ee902f65e7daa34a20c0d4e65944352bc5311b6a96cb47ef"} Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.030538 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a95c9a423b951740ee902f65e7daa34a20c0d4e65944352bc5311b6a96cb47ef" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.030627 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.030637 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.030681 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wc9k9" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.057557 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.057587 4876 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.057598 4876 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.057609 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwlrc\" (UniqueName: \"kubernetes.io/projected/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-kube-api-access-hwlrc\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.057619 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.057627 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.077760 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.078751 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.888221 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-577c476758-ksnrz"] Dec 05 08:36:17 crc kubenswrapper[4876]: E1205 08:36:17.888783 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" containerName="keystone-bootstrap" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.888803 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" containerName="keystone-bootstrap" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.888994 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" containerName="keystone-bootstrap" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.889544 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.892681 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.892688 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fmb5b" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.893796 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.894021 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.894177 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.895151 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 05 08:36:17 crc kubenswrapper[4876]: I1205 08:36:17.908618 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-577c476758-ksnrz"] Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.075048 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-internal-tls-certs\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.075111 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-config-data\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.075158 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm8sv\" (UniqueName: \"kubernetes.io/projected/5612eb81-67eb-4559-b21a-549e4becca72-kube-api-access-xm8sv\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.075202 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-scripts\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.075217 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-public-tls-certs\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.075261 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-credential-keys\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.075284 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-fernet-keys\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.075300 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-combined-ca-bundle\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.177409 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-credential-keys\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.177458 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-fernet-keys\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.177473 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-combined-ca-bundle\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.177504 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-internal-tls-certs\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.177538 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-config-data\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.177585 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm8sv\" (UniqueName: \"kubernetes.io/projected/5612eb81-67eb-4559-b21a-549e4becca72-kube-api-access-xm8sv\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.177677 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-public-tls-certs\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.177720 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-scripts\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.185741 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-scripts\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.185862 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-credential-keys\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.186366 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-internal-tls-certs\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.186975 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-combined-ca-bundle\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.188424 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-config-data\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.191399 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-fernet-keys\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.193649 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5612eb81-67eb-4559-b21a-549e4becca72-public-tls-certs\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.197338 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm8sv\" (UniqueName: \"kubernetes.io/projected/5612eb81-67eb-4559-b21a-549e4becca72-kube-api-access-xm8sv\") pod \"keystone-577c476758-ksnrz\" (UID: \"5612eb81-67eb-4559-b21a-549e4becca72\") " pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.266029 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.415077 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.415138 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.450545 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:18 crc kubenswrapper[4876]: I1205 08:36:18.465533 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.045859 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.046024 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.255129 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.314218 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-hrms7"] Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.316217 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" podUID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" containerName="dnsmasq-dns" containerID="cri-o://a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b" gracePeriod=10 Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.630071 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n5bmx" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.636615 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-m2khz" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.717483 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-scripts\") pod \"460f93bd-e732-42c0-8e75-625f3346800c\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.717645 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7k8q\" (UniqueName: \"kubernetes.io/projected/460f93bd-e732-42c0-8e75-625f3346800c-kube-api-access-q7k8q\") pod \"460f93bd-e732-42c0-8e75-625f3346800c\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.717663 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-config-data\") pod \"460f93bd-e732-42c0-8e75-625f3346800c\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.717710 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-combined-ca-bundle\") pod \"460f93bd-e732-42c0-8e75-625f3346800c\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.717812 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/460f93bd-e732-42c0-8e75-625f3346800c-logs\") pod \"460f93bd-e732-42c0-8e75-625f3346800c\" (UID: \"460f93bd-e732-42c0-8e75-625f3346800c\") " Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.718540 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/460f93bd-e732-42c0-8e75-625f3346800c-logs" (OuterVolumeSpecName: "logs") pod "460f93bd-e732-42c0-8e75-625f3346800c" (UID: "460f93bd-e732-42c0-8e75-625f3346800c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.744403 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-scripts" (OuterVolumeSpecName: "scripts") pod "460f93bd-e732-42c0-8e75-625f3346800c" (UID: "460f93bd-e732-42c0-8e75-625f3346800c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.760331 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/460f93bd-e732-42c0-8e75-625f3346800c-kube-api-access-q7k8q" (OuterVolumeSpecName: "kube-api-access-q7k8q") pod "460f93bd-e732-42c0-8e75-625f3346800c" (UID: "460f93bd-e732-42c0-8e75-625f3346800c"). InnerVolumeSpecName "kube-api-access-q7k8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.783510 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-config-data" (OuterVolumeSpecName: "config-data") pod "460f93bd-e732-42c0-8e75-625f3346800c" (UID: "460f93bd-e732-42c0-8e75-625f3346800c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.802115 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "460f93bd-e732-42c0-8e75-625f3346800c" (UID: "460f93bd-e732-42c0-8e75-625f3346800c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.819616 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-db-sync-config-data\") pod \"2d18b829-e516-41de-989d-bea32b940409\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.819734 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-combined-ca-bundle\") pod \"2d18b829-e516-41de-989d-bea32b940409\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.819778 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnv8k\" (UniqueName: \"kubernetes.io/projected/2d18b829-e516-41de-989d-bea32b940409-kube-api-access-hnv8k\") pod \"2d18b829-e516-41de-989d-bea32b940409\" (UID: \"2d18b829-e516-41de-989d-bea32b940409\") " Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.820258 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/460f93bd-e732-42c0-8e75-625f3346800c-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.820278 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.820287 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7k8q\" (UniqueName: \"kubernetes.io/projected/460f93bd-e732-42c0-8e75-625f3346800c-kube-api-access-q7k8q\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.820295 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.820304 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460f93bd-e732-42c0-8e75-625f3346800c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.826279 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d18b829-e516-41de-989d-bea32b940409-kube-api-access-hnv8k" (OuterVolumeSpecName: "kube-api-access-hnv8k") pod "2d18b829-e516-41de-989d-bea32b940409" (UID: "2d18b829-e516-41de-989d-bea32b940409"). InnerVolumeSpecName "kube-api-access-hnv8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.839273 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2d18b829-e516-41de-989d-bea32b940409" (UID: "2d18b829-e516-41de-989d-bea32b940409"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.890529 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d18b829-e516-41de-989d-bea32b940409" (UID: "2d18b829-e516-41de-989d-bea32b940409"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.928544 4876 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.928569 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d18b829-e516-41de-989d-bea32b940409-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.928579 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnv8k\" (UniqueName: \"kubernetes.io/projected/2d18b829-e516-41de-989d-bea32b940409-kube-api-access-hnv8k\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:19 crc kubenswrapper[4876]: I1205 08:36:19.948843 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.071936 4876 generic.go:334] "Generic (PLEG): container finished" podID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" containerID="a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b" exitCode=0 Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.072015 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" event={"ID":"fd5f73d3-2859-4e4d-a5d8-02acd110f11a","Type":"ContainerDied","Data":"a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b"} Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.072043 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" event={"ID":"fd5f73d3-2859-4e4d-a5d8-02acd110f11a","Type":"ContainerDied","Data":"8e0033ba63663d8341cbe3946db8b436b6e7ff35b314ec1b52fff083b7049224"} Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.072079 4876 scope.go:117] "RemoveContainer" containerID="a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.072239 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-hrms7" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.081415 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n5bmx" event={"ID":"460f93bd-e732-42c0-8e75-625f3346800c","Type":"ContainerDied","Data":"0711a02349f25779b80825587a7c137737de21cdae5259c054ad030ce613802c"} Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.081450 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0711a02349f25779b80825587a7c137737de21cdae5259c054ad030ce613802c" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.081504 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n5bmx" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.085363 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-m2khz" event={"ID":"2d18b829-e516-41de-989d-bea32b940409","Type":"ContainerDied","Data":"43edfbacf605443d0d0944225e3d6de9f7ea2f4b7fb9c285098cf479b068b686"} Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.085397 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43edfbacf605443d0d0944225e3d6de9f7ea2f4b7fb9c285098cf479b068b686" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.085447 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-m2khz" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.095662 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ff8f8c8c7-rx6w7" event={"ID":"cd9bce2b-4ecb-45b0-a1b2-309c3f450a47","Type":"ContainerStarted","Data":"6e8944b087183e95b5b32d77624bbb26c0ee39e819169c5778982fb465dc3ea1"} Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.095797 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.107357 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5","Type":"ContainerStarted","Data":"f10cb06c492843f0aa303245e23e0d29c816b8c13819ee1f6780e7985654a987"} Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.130429 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvf6d\" (UniqueName: \"kubernetes.io/projected/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-kube-api-access-wvf6d\") pod \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.130574 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-swift-storage-0\") pod \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.130602 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-nb\") pod \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.131154 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-sb\") pod \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.131240 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-svc\") pod \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.131274 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-config\") pod \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\" (UID: \"fd5f73d3-2859-4e4d-a5d8-02acd110f11a\") " Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.132036 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5ff8f8c8c7-rx6w7" podStartSLOduration=9.132019579 podStartE2EDuration="9.132019579s" podCreationTimestamp="2025-12-05 08:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:20.119533602 +0000 UTC m=+1244.608198234" watchObservedRunningTime="2025-12-05 08:36:20.132019579 +0000 UTC m=+1244.620684191" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.142069 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-kube-api-access-wvf6d" (OuterVolumeSpecName: "kube-api-access-wvf6d") pod "fd5f73d3-2859-4e4d-a5d8-02acd110f11a" (UID: "fd5f73d3-2859-4e4d-a5d8-02acd110f11a"). InnerVolumeSpecName "kube-api-access-wvf6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.169246 4876 scope.go:117] "RemoveContainer" containerID="1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.179037 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fd5f73d3-2859-4e4d-a5d8-02acd110f11a" (UID: "fd5f73d3-2859-4e4d-a5d8-02acd110f11a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.184798 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fd5f73d3-2859-4e4d-a5d8-02acd110f11a" (UID: "fd5f73d3-2859-4e4d-a5d8-02acd110f11a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.186243 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fd5f73d3-2859-4e4d-a5d8-02acd110f11a" (UID: "fd5f73d3-2859-4e4d-a5d8-02acd110f11a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.188702 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fd5f73d3-2859-4e4d-a5d8-02acd110f11a" (UID: "fd5f73d3-2859-4e4d-a5d8-02acd110f11a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.193964 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-config" (OuterVolumeSpecName: "config") pod "fd5f73d3-2859-4e4d-a5d8-02acd110f11a" (UID: "fd5f73d3-2859-4e4d-a5d8-02acd110f11a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.205022 4876 scope.go:117] "RemoveContainer" containerID="a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b" Dec 05 08:36:20 crc kubenswrapper[4876]: E1205 08:36:20.211649 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b\": container with ID starting with a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b not found: ID does not exist" containerID="a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.212657 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b"} err="failed to get container status \"a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b\": rpc error: code = NotFound desc = could not find container \"a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b\": container with ID starting with a15a05b343b72e0b2257875db9ebf9126ec030e5361cf1dd4f074d8aadfa0a6b not found: ID does not exist" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.212709 4876 scope.go:117] "RemoveContainer" containerID="1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d" Dec 05 08:36:20 crc kubenswrapper[4876]: E1205 08:36:20.213228 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d\": container with ID starting with 1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d not found: ID does not exist" containerID="1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.213269 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d"} err="failed to get container status \"1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d\": rpc error: code = NotFound desc = could not find container \"1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d\": container with ID starting with 1bbef83f569e6b5c83f70e2c16429ab17bd87ed6980b42de6f49f4e4188daf1d not found: ID does not exist" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.235289 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.235331 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.235363 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.235377 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvf6d\" (UniqueName: \"kubernetes.io/projected/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-kube-api-access-wvf6d\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.235390 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.235397 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd5f73d3-2859-4e4d-a5d8-02acd110f11a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.255775 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-577c476758-ksnrz"] Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.404212 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-hrms7"] Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.410825 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-hrms7"] Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.749621 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-56d57865bd-mfljr"] Dec 05 08:36:20 crc kubenswrapper[4876]: E1205 08:36:20.750361 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d18b829-e516-41de-989d-bea32b940409" containerName="barbican-db-sync" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.750385 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d18b829-e516-41de-989d-bea32b940409" containerName="barbican-db-sync" Dec 05 08:36:20 crc kubenswrapper[4876]: E1205 08:36:20.750412 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="460f93bd-e732-42c0-8e75-625f3346800c" containerName="placement-db-sync" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.750421 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="460f93bd-e732-42c0-8e75-625f3346800c" containerName="placement-db-sync" Dec 05 08:36:20 crc kubenswrapper[4876]: E1205 08:36:20.750445 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" containerName="init" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.750453 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" containerName="init" Dec 05 08:36:20 crc kubenswrapper[4876]: E1205 08:36:20.750466 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" containerName="dnsmasq-dns" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.750474 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" containerName="dnsmasq-dns" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.750654 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" containerName="dnsmasq-dns" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.750669 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d18b829-e516-41de-989d-bea32b940409" containerName="barbican-db-sync" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.750679 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="460f93bd-e732-42c0-8e75-625f3346800c" containerName="placement-db-sync" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.751671 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.754533 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.754601 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-djmxn" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.754760 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.754801 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.754814 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.762377 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-56d57865bd-mfljr"] Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.846997 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-public-tls-certs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.847333 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-internal-tls-certs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.847369 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-scripts\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.847394 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27487f6b-feb8-47c1-98db-5ed28120b3f3-logs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.847438 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-combined-ca-bundle\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.847668 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-config-data\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.847762 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbgcz\" (UniqueName: \"kubernetes.io/projected/27487f6b-feb8-47c1-98db-5ed28120b3f3-kube-api-access-lbgcz\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.949510 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-config-data\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.949595 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbgcz\" (UniqueName: \"kubernetes.io/projected/27487f6b-feb8-47c1-98db-5ed28120b3f3-kube-api-access-lbgcz\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.949643 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-public-tls-certs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.949705 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-internal-tls-certs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.949741 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-scripts\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.949767 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27487f6b-feb8-47c1-98db-5ed28120b3f3-logs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.949819 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-combined-ca-bundle\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.957322 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-config-data\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.960292 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27487f6b-feb8-47c1-98db-5ed28120b3f3-logs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.963005 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-internal-tls-certs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:20 crc kubenswrapper[4876]: I1205 08:36:20.994943 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-scripts\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:20.969657 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-combined-ca-bundle\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.025369 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbgcz\" (UniqueName: \"kubernetes.io/projected/27487f6b-feb8-47c1-98db-5ed28120b3f3-kube-api-access-lbgcz\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.025421 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27487f6b-feb8-47c1-98db-5ed28120b3f3-public-tls-certs\") pod \"placement-56d57865bd-mfljr\" (UID: \"27487f6b-feb8-47c1-98db-5ed28120b3f3\") " pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.053966 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6f66c6dbdf-59mzz"] Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.080590 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.082710 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.105090 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mlf96" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.105318 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-69575954c8-dwfz7"] Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.105514 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.106191 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.125398 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.158234 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6f66c6dbdf-59mzz"] Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.160232 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-config-data-custom\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.160307 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa8ee581-b9fc-4c2d-8139-b3d827f71697-logs\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.160332 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-combined-ca-bundle\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.160368 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-config-data\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.160417 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djxmc\" (UniqueName: \"kubernetes.io/projected/aa8ee581-b9fc-4c2d-8139-b3d827f71697-kube-api-access-djxmc\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.167419 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.170634 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vxcln" event={"ID":"62f1d7bc-1a36-4fc0-a94d-72fcedda052e","Type":"ContainerStarted","Data":"5b43864474344dca68358c926da69812978879965307a50d4093bbef607f1897"} Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.190266 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-s2gmz"] Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.197775 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.199534 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-69575954c8-dwfz7"] Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.209970 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-s2gmz"] Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.217774 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xbjb2" event={"ID":"5a1ae752-b331-4d0f-8039-e5720312f674","Type":"ContainerStarted","Data":"215e4e1b05c2f0f608af1bb1647a803db99e077dabcf6a941b8d8a1170b223fc"} Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.254549 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-577c476758-ksnrz" event={"ID":"5612eb81-67eb-4559-b21a-549e4becca72","Type":"ContainerStarted","Data":"04268a6d0af554a865d59b4f2772d9643412c5c85f70bb6f5adbfe02bdc04de5"} Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.254594 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.254604 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-577c476758-ksnrz" event={"ID":"5612eb81-67eb-4559-b21a-549e4becca72","Type":"ContainerStarted","Data":"2afeb214eac9fec52093144717a1001f2e8a93f88a32980c5ebfc5eb5bd1231d"} Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg6wd\" (UniqueName: \"kubernetes.io/projected/23f75a46-ea18-4217-bcb8-26d5652f95eb-kube-api-access-cg6wd\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263784 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa8ee581-b9fc-4c2d-8139-b3d827f71697-logs\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263808 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-config\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263826 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263861 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-combined-ca-bundle\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263914 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263933 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-config-data\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263952 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-combined-ca-bundle\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.263999 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-config-data\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.264031 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4lz2\" (UniqueName: \"kubernetes.io/projected/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-kube-api-access-z4lz2\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.264048 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-svc\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.264072 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.264120 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djxmc\" (UniqueName: \"kubernetes.io/projected/aa8ee581-b9fc-4c2d-8139-b3d827f71697-kube-api-access-djxmc\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.264175 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-config-data-custom\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.264210 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23f75a46-ea18-4217-bcb8-26d5652f95eb-logs\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.264237 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-config-data-custom\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.268050 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7c5857d458-hs895"] Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.268630 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa8ee581-b9fc-4c2d-8139-b3d827f71697-logs\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.269591 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.275132 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.284400 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-config-data\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.289444 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-combined-ca-bundle\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.296004 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aa8ee581-b9fc-4c2d-8139-b3d827f71697-config-data-custom\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.314173 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c5857d458-hs895"] Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.318695 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djxmc\" (UniqueName: \"kubernetes.io/projected/aa8ee581-b9fc-4c2d-8139-b3d827f71697-kube-api-access-djxmc\") pod \"barbican-worker-6f66c6dbdf-59mzz\" (UID: \"aa8ee581-b9fc-4c2d-8139-b3d827f71697\") " pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.326002 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vxcln" podStartSLOduration=3.894529134 podStartE2EDuration="41.325984205s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="2025-12-05 08:35:42.296581824 +0000 UTC m=+1206.785246446" lastFinishedPulling="2025-12-05 08:36:19.728036895 +0000 UTC m=+1244.216701517" observedRunningTime="2025-12-05 08:36:21.217768949 +0000 UTC m=+1245.706433571" watchObservedRunningTime="2025-12-05 08:36:21.325984205 +0000 UTC m=+1245.814648827" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.356831 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-xbjb2" podStartSLOduration=3.924825611 podStartE2EDuration="41.356810195s" podCreationTimestamp="2025-12-05 08:35:40 +0000 UTC" firstStartedPulling="2025-12-05 08:35:42.252158209 +0000 UTC m=+1206.740822831" lastFinishedPulling="2025-12-05 08:36:19.684142793 +0000 UTC m=+1244.172807415" observedRunningTime="2025-12-05 08:36:21.265281314 +0000 UTC m=+1245.753945936" watchObservedRunningTime="2025-12-05 08:36:21.356810195 +0000 UTC m=+1245.845474817" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367516 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d63901-e276-47e4-b544-5763ff670a20-logs\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367561 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxcdh\" (UniqueName: \"kubernetes.io/projected/53d63901-e276-47e4-b544-5763ff670a20-kube-api-access-jxcdh\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367641 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4lz2\" (UniqueName: \"kubernetes.io/projected/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-kube-api-access-z4lz2\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367662 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-svc\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367690 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367716 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data-custom\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367833 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23f75a46-ea18-4217-bcb8-26d5652f95eb-logs\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367852 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-combined-ca-bundle\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367892 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-config-data-custom\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.367960 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg6wd\" (UniqueName: \"kubernetes.io/projected/23f75a46-ea18-4217-bcb8-26d5652f95eb-kube-api-access-cg6wd\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.368022 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-config\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.368043 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.368071 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.368088 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.368108 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-config-data\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.368133 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-combined-ca-bundle\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.368819 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-svc\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.369122 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-config\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.369402 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23f75a46-ea18-4217-bcb8-26d5652f95eb-logs\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.369645 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.370859 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.373390 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.377287 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-577c476758-ksnrz" podStartSLOduration=4.377263418 podStartE2EDuration="4.377263418s" podCreationTimestamp="2025-12-05 08:36:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:21.302251649 +0000 UTC m=+1245.790916271" watchObservedRunningTime="2025-12-05 08:36:21.377263418 +0000 UTC m=+1245.865928040" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.379987 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-config-data-custom\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.386433 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-config-data\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.392564 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23f75a46-ea18-4217-bcb8-26d5652f95eb-combined-ca-bundle\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.394921 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg6wd\" (UniqueName: \"kubernetes.io/projected/23f75a46-ea18-4217-bcb8-26d5652f95eb-kube-api-access-cg6wd\") pod \"barbican-keystone-listener-69575954c8-dwfz7\" (UID: \"23f75a46-ea18-4217-bcb8-26d5652f95eb\") " pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.398574 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4lz2\" (UniqueName: \"kubernetes.io/projected/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-kube-api-access-z4lz2\") pod \"dnsmasq-dns-688c87cc99-s2gmz\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.469474 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.469734 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d63901-e276-47e4-b544-5763ff670a20-logs\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.469753 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxcdh\" (UniqueName: \"kubernetes.io/projected/53d63901-e276-47e4-b544-5763ff670a20-kube-api-access-jxcdh\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.469796 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data-custom\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.469858 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-combined-ca-bundle\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.470640 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d63901-e276-47e4-b544-5763ff670a20-logs\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.474232 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-combined-ca-bundle\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.476338 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data-custom\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.477142 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.477603 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6f66c6dbdf-59mzz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.486475 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxcdh\" (UniqueName: \"kubernetes.io/projected/53d63901-e276-47e4-b544-5763ff670a20-kube-api-access-jxcdh\") pod \"barbican-api-7c5857d458-hs895\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.492471 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.543588 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.645455 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.806911 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-56d57865bd-mfljr"] Dec 05 08:36:21 crc kubenswrapper[4876]: W1205 08:36:21.860477 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27487f6b_feb8_47c1_98db_5ed28120b3f3.slice/crio-46623e29491428e373e7ccb1a6dd54dceae39566f13d438bb305d968bf23ef1e WatchSource:0}: Error finding container 46623e29491428e373e7ccb1a6dd54dceae39566f13d438bb305d968bf23ef1e: Status 404 returned error can't find the container with id 46623e29491428e373e7ccb1a6dd54dceae39566f13d438bb305d968bf23ef1e Dec 05 08:36:21 crc kubenswrapper[4876]: I1205 08:36:21.871782 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd5f73d3-2859-4e4d-a5d8-02acd110f11a" path="/var/lib/kubelet/pods/fd5f73d3-2859-4e4d-a5d8-02acd110f11a/volumes" Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.189166 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6f66c6dbdf-59mzz"] Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.275169 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-s2gmz"] Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.279808 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f66c6dbdf-59mzz" event={"ID":"aa8ee581-b9fc-4c2d-8139-b3d827f71697","Type":"ContainerStarted","Data":"eb5cbeb87d4272fc8dce6ea16e56cb1bb50550c455103225e1bf27c267e55356"} Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.284367 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-69575954c8-dwfz7"] Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.289457 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56d57865bd-mfljr" event={"ID":"27487f6b-feb8-47c1-98db-5ed28120b3f3","Type":"ContainerStarted","Data":"cbc8a83bbf1d02530f86f0cc5c5e64426022a89cb0d973bf47dad188590c519f"} Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.289494 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56d57865bd-mfljr" event={"ID":"27487f6b-feb8-47c1-98db-5ed28120b3f3","Type":"ContainerStarted","Data":"46623e29491428e373e7ccb1a6dd54dceae39566f13d438bb305d968bf23ef1e"} Dec 05 08:36:22 crc kubenswrapper[4876]: W1205 08:36:22.296460 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23f75a46_ea18_4217_bcb8_26d5652f95eb.slice/crio-5c103864d6b7562f53618cfea300262423de84d2ae0045096bcdfb073def8e78 WatchSource:0}: Error finding container 5c103864d6b7562f53618cfea300262423de84d2ae0045096bcdfb073def8e78: Status 404 returned error can't find the container with id 5c103864d6b7562f53618cfea300262423de84d2ae0045096bcdfb073def8e78 Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.429828 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.430373 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 08:36:22 crc kubenswrapper[4876]: I1205 08:36:22.451214 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7c5857d458-hs895"] Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.309740 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-56d57865bd-mfljr" event={"ID":"27487f6b-feb8-47c1-98db-5ed28120b3f3","Type":"ContainerStarted","Data":"b5bed8dd12908a0a17472d3ebcd1c6bef2d8274e52715ce502182307f78c8d26"} Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.310418 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.310548 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.316504 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" event={"ID":"23f75a46-ea18-4217-bcb8-26d5652f95eb","Type":"ContainerStarted","Data":"5c103864d6b7562f53618cfea300262423de84d2ae0045096bcdfb073def8e78"} Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.325227 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c5857d458-hs895" event={"ID":"53d63901-e276-47e4-b544-5763ff670a20","Type":"ContainerStarted","Data":"735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d"} Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.325271 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c5857d458-hs895" event={"ID":"53d63901-e276-47e4-b544-5763ff670a20","Type":"ContainerStarted","Data":"ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7"} Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.325285 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c5857d458-hs895" event={"ID":"53d63901-e276-47e4-b544-5763ff670a20","Type":"ContainerStarted","Data":"380ec081a95e02b3a178730d9e5691f2940c131360e1e0c7a01b5542cc34dbc9"} Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.326057 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.326079 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.334195 4876 generic.go:334] "Generic (PLEG): container finished" podID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerID="9374100375ae0cceb7ea0ab64339a9eabe2188c59fc65a757ab65efa8242303e" exitCode=0 Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.335525 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" event={"ID":"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c","Type":"ContainerDied","Data":"9374100375ae0cceb7ea0ab64339a9eabe2188c59fc65a757ab65efa8242303e"} Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.335551 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" event={"ID":"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c","Type":"ContainerStarted","Data":"269d8b3bc022898adcb8a965c1e0c0601da0d1d4b376b1a503e0677210857c81"} Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.336894 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-56d57865bd-mfljr" podStartSLOduration=3.336870296 podStartE2EDuration="3.336870296s" podCreationTimestamp="2025-12-05 08:36:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:23.329444824 +0000 UTC m=+1247.818109446" watchObservedRunningTime="2025-12-05 08:36:23.336870296 +0000 UTC m=+1247.825534918" Dec 05 08:36:23 crc kubenswrapper[4876]: I1205 08:36:23.372675 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7c5857d458-hs895" podStartSLOduration=2.372657837 podStartE2EDuration="2.372657837s" podCreationTimestamp="2025-12-05 08:36:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:23.355387184 +0000 UTC m=+1247.844051826" watchObservedRunningTime="2025-12-05 08:36:23.372657837 +0000 UTC m=+1247.861322459" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.409054 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9d7c76978-sl6zr"] Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.411371 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.432113 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.432182 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.452062 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9d7c76978-sl6zr"] Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.571087 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-config-data\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.571379 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-logs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.571436 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-config-data-custom\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.571631 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-combined-ca-bundle\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.571659 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jlw9\" (UniqueName: \"kubernetes.io/projected/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-kube-api-access-5jlw9\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.571697 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-public-tls-certs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.572874 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-internal-tls-certs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.680656 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-combined-ca-bundle\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.680725 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jlw9\" (UniqueName: \"kubernetes.io/projected/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-kube-api-access-5jlw9\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.680764 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-public-tls-certs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.680842 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-internal-tls-certs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.680872 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-config-data\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.680921 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-logs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.680959 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-config-data-custom\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.683053 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-logs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.689257 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-combined-ca-bundle\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.690572 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-config-data-custom\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.691737 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-internal-tls-certs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.694867 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-config-data\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.704019 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-public-tls-certs\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.707690 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jlw9\" (UniqueName: \"kubernetes.io/projected/d9c4abfe-e5d6-4a04-88ff-3e721675a75e-kube-api-access-5jlw9\") pod \"barbican-api-9d7c76978-sl6zr\" (UID: \"d9c4abfe-e5d6-4a04-88ff-3e721675a75e\") " pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:24 crc kubenswrapper[4876]: I1205 08:36:24.836524 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.314640 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9d7c76978-sl6zr"] Dec 05 08:36:25 crc kubenswrapper[4876]: W1205 08:36:25.325720 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9c4abfe_e5d6_4a04_88ff_3e721675a75e.slice/crio-2cd5f437b49ddf860242b3bc770655e31a1cc787fc1e2fdef481fa2ccf910026 WatchSource:0}: Error finding container 2cd5f437b49ddf860242b3bc770655e31a1cc787fc1e2fdef481fa2ccf910026: Status 404 returned error can't find the container with id 2cd5f437b49ddf860242b3bc770655e31a1cc787fc1e2fdef481fa2ccf910026 Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.456261 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f66c6dbdf-59mzz" event={"ID":"aa8ee581-b9fc-4c2d-8139-b3d827f71697","Type":"ContainerStarted","Data":"c2818efb17f47224b6684dc8cc77225500cf502206caa920b1a3da8cea16c616"} Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.456326 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f66c6dbdf-59mzz" event={"ID":"aa8ee581-b9fc-4c2d-8139-b3d827f71697","Type":"ContainerStarted","Data":"4be31e976a56ac87c96411f1324a900cb70858371a437fdb7fc569fd7243abae"} Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.470369 4876 generic.go:334] "Generic (PLEG): container finished" podID="5a1ae752-b331-4d0f-8039-e5720312f674" containerID="215e4e1b05c2f0f608af1bb1647a803db99e077dabcf6a941b8d8a1170b223fc" exitCode=0 Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.470446 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xbjb2" event={"ID":"5a1ae752-b331-4d0f-8039-e5720312f674","Type":"ContainerDied","Data":"215e4e1b05c2f0f608af1bb1647a803db99e077dabcf6a941b8d8a1170b223fc"} Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.477974 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9d7c76978-sl6zr" event={"ID":"d9c4abfe-e5d6-4a04-88ff-3e721675a75e","Type":"ContainerStarted","Data":"2cd5f437b49ddf860242b3bc770655e31a1cc787fc1e2fdef481fa2ccf910026"} Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.479856 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6f66c6dbdf-59mzz" podStartSLOduration=3.472985807 podStartE2EDuration="5.479842213s" podCreationTimestamp="2025-12-05 08:36:20 +0000 UTC" firstStartedPulling="2025-12-05 08:36:22.214801169 +0000 UTC m=+1246.703465791" lastFinishedPulling="2025-12-05 08:36:24.221657575 +0000 UTC m=+1248.710322197" observedRunningTime="2025-12-05 08:36:25.474987185 +0000 UTC m=+1249.963651817" watchObservedRunningTime="2025-12-05 08:36:25.479842213 +0000 UTC m=+1249.968506835" Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.492111 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" event={"ID":"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c","Type":"ContainerStarted","Data":"19c25aedd431a59ef04a19e058aa9006fbc8142b46d2cb97b8db5944e08b6a96"} Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.492170 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.501373 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" event={"ID":"23f75a46-ea18-4217-bcb8-26d5652f95eb","Type":"ContainerStarted","Data":"d3f442bf9c362106a87a219f9234bca38b539c034d67ba80a160900e59c4aeed"} Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.501661 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" event={"ID":"23f75a46-ea18-4217-bcb8-26d5652f95eb","Type":"ContainerStarted","Data":"7880531948c55c62936d06b692b59e599c26e9308e1bf1ef959abd27a0304ecd"} Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.521237 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" podStartSLOduration=5.5212133439999995 podStartE2EDuration="5.521213344s" podCreationTimestamp="2025-12-05 08:36:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:25.519204226 +0000 UTC m=+1250.007868848" watchObservedRunningTime="2025-12-05 08:36:25.521213344 +0000 UTC m=+1250.009877976" Dec 05 08:36:25 crc kubenswrapper[4876]: I1205 08:36:25.554106 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-69575954c8-dwfz7" podStartSLOduration=3.617894941 podStartE2EDuration="5.554085451s" podCreationTimestamp="2025-12-05 08:36:20 +0000 UTC" firstStartedPulling="2025-12-05 08:36:22.30703497 +0000 UTC m=+1246.795699592" lastFinishedPulling="2025-12-05 08:36:24.24322548 +0000 UTC m=+1248.731890102" observedRunningTime="2025-12-05 08:36:25.544744665 +0000 UTC m=+1250.033409287" watchObservedRunningTime="2025-12-05 08:36:25.554085451 +0000 UTC m=+1250.042750073" Dec 05 08:36:26 crc kubenswrapper[4876]: I1205 08:36:26.511759 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9d7c76978-sl6zr" event={"ID":"d9c4abfe-e5d6-4a04-88ff-3e721675a75e","Type":"ContainerStarted","Data":"bbaa8a77dc013f76e44459202dc72af0cffb4af22a9a8179b6a6ff8a3d26ebd5"} Dec 05 08:36:26 crc kubenswrapper[4876]: I1205 08:36:26.512590 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9d7c76978-sl6zr" event={"ID":"d9c4abfe-e5d6-4a04-88ff-3e721675a75e","Type":"ContainerStarted","Data":"d13de72578e504764793b3f3ae89892a448430f5b17f764b123bd1bd3e0b0bfa"} Dec 05 08:36:26 crc kubenswrapper[4876]: I1205 08:36:26.541217 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9d7c76978-sl6zr" podStartSLOduration=2.5411985489999998 podStartE2EDuration="2.541198549s" podCreationTimestamp="2025-12-05 08:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:26.531223734 +0000 UTC m=+1251.019888366" watchObservedRunningTime="2025-12-05 08:36:26.541198549 +0000 UTC m=+1251.029863171" Dec 05 08:36:27 crc kubenswrapper[4876]: I1205 08:36:27.521952 4876 generic.go:334] "Generic (PLEG): container finished" podID="62f1d7bc-1a36-4fc0-a94d-72fcedda052e" containerID="5b43864474344dca68358c926da69812978879965307a50d4093bbef607f1897" exitCode=0 Dec 05 08:36:27 crc kubenswrapper[4876]: I1205 08:36:27.522183 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vxcln" event={"ID":"62f1d7bc-1a36-4fc0-a94d-72fcedda052e","Type":"ContainerDied","Data":"5b43864474344dca68358c926da69812978879965307a50d4093bbef607f1897"} Dec 05 08:36:27 crc kubenswrapper[4876]: I1205 08:36:27.522622 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:27 crc kubenswrapper[4876]: I1205 08:36:27.522707 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.294495 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xbjb2" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.471957 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nptt\" (UniqueName: \"kubernetes.io/projected/5a1ae752-b331-4d0f-8039-e5720312f674-kube-api-access-2nptt\") pod \"5a1ae752-b331-4d0f-8039-e5720312f674\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.472226 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-config-data\") pod \"5a1ae752-b331-4d0f-8039-e5720312f674\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.472304 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-combined-ca-bundle\") pod \"5a1ae752-b331-4d0f-8039-e5720312f674\" (UID: \"5a1ae752-b331-4d0f-8039-e5720312f674\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.478885 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1ae752-b331-4d0f-8039-e5720312f674-kube-api-access-2nptt" (OuterVolumeSpecName: "kube-api-access-2nptt") pod "5a1ae752-b331-4d0f-8039-e5720312f674" (UID: "5a1ae752-b331-4d0f-8039-e5720312f674"). InnerVolumeSpecName "kube-api-access-2nptt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.501369 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a1ae752-b331-4d0f-8039-e5720312f674" (UID: "5a1ae752-b331-4d0f-8039-e5720312f674"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.543963 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xbjb2" event={"ID":"5a1ae752-b331-4d0f-8039-e5720312f674","Type":"ContainerDied","Data":"de2e048dbe1393e79577008e7b06a819b08783e180454a2f426b1caa25bcee00"} Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.544001 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de2e048dbe1393e79577008e7b06a819b08783e180454a2f426b1caa25bcee00" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.544038 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xbjb2" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.550266 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-config-data" (OuterVolumeSpecName: "config-data") pod "5a1ae752-b331-4d0f-8039-e5720312f674" (UID: "5a1ae752-b331-4d0f-8039-e5720312f674"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.574508 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.574542 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a1ae752-b331-4d0f-8039-e5720312f674-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.574553 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nptt\" (UniqueName: \"kubernetes.io/projected/5a1ae752-b331-4d0f-8039-e5720312f674-kube-api-access-2nptt\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.740434 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vxcln" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.888467 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55btw\" (UniqueName: \"kubernetes.io/projected/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-kube-api-access-55btw\") pod \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.888595 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-config-data\") pod \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.888712 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-combined-ca-bundle\") pod \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.888931 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-scripts\") pod \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.889007 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-etc-machine-id\") pod \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.889191 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-db-sync-config-data\") pod \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\" (UID: \"62f1d7bc-1a36-4fc0-a94d-72fcedda052e\") " Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.889220 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "62f1d7bc-1a36-4fc0-a94d-72fcedda052e" (UID: "62f1d7bc-1a36-4fc0-a94d-72fcedda052e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.890136 4876 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.892719 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "62f1d7bc-1a36-4fc0-a94d-72fcedda052e" (UID: "62f1d7bc-1a36-4fc0-a94d-72fcedda052e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.892718 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-kube-api-access-55btw" (OuterVolumeSpecName: "kube-api-access-55btw") pod "62f1d7bc-1a36-4fc0-a94d-72fcedda052e" (UID: "62f1d7bc-1a36-4fc0-a94d-72fcedda052e"). InnerVolumeSpecName "kube-api-access-55btw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.894180 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-scripts" (OuterVolumeSpecName: "scripts") pod "62f1d7bc-1a36-4fc0-a94d-72fcedda052e" (UID: "62f1d7bc-1a36-4fc0-a94d-72fcedda052e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.920582 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62f1d7bc-1a36-4fc0-a94d-72fcedda052e" (UID: "62f1d7bc-1a36-4fc0-a94d-72fcedda052e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.970673 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-config-data" (OuterVolumeSpecName: "config-data") pod "62f1d7bc-1a36-4fc0-a94d-72fcedda052e" (UID: "62f1d7bc-1a36-4fc0-a94d-72fcedda052e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:29 crc kubenswrapper[4876]: E1205 08:36:29.983642 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.992125 4876 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.992174 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55btw\" (UniqueName: \"kubernetes.io/projected/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-kube-api-access-55btw\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.992209 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.992229 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:29 crc kubenswrapper[4876]: I1205 08:36:29.992246 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62f1d7bc-1a36-4fc0-a94d-72fcedda052e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:30 crc kubenswrapper[4876]: I1205 08:36:30.556995 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vxcln" Dec 05 08:36:30 crc kubenswrapper[4876]: I1205 08:36:30.556975 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vxcln" event={"ID":"62f1d7bc-1a36-4fc0-a94d-72fcedda052e","Type":"ContainerDied","Data":"2988e26a23f24514afda91af74167f2c6be3667fe48867158bb1228754988893"} Dec 05 08:36:30 crc kubenswrapper[4876]: I1205 08:36:30.557155 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2988e26a23f24514afda91af74167f2c6be3667fe48867158bb1228754988893" Dec 05 08:36:30 crc kubenswrapper[4876]: I1205 08:36:30.560099 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5","Type":"ContainerStarted","Data":"a3451ab3ae1050cd70ba23c95cba113de1a07431dd1e3918f3f668bef54c4cd9"} Dec 05 08:36:30 crc kubenswrapper[4876]: I1205 08:36:30.560262 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="ceilometer-notification-agent" containerID="cri-o://bdb2749c0bdbf1d92e5bea8bf5cb1424c07fab5d7e2895e4bfcb0f9ab6f0c4d8" gracePeriod=30 Dec 05 08:36:30 crc kubenswrapper[4876]: I1205 08:36:30.560348 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="proxy-httpd" containerID="cri-o://a3451ab3ae1050cd70ba23c95cba113de1a07431dd1e3918f3f668bef54c4cd9" gracePeriod=30 Dec 05 08:36:30 crc kubenswrapper[4876]: I1205 08:36:30.560278 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:36:30 crc kubenswrapper[4876]: I1205 08:36:30.560366 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="sg-core" containerID="cri-o://f10cb06c492843f0aa303245e23e0d29c816b8c13819ee1f6780e7985654a987" gracePeriod=30 Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.054775 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:31 crc kubenswrapper[4876]: E1205 08:36:31.055653 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1ae752-b331-4d0f-8039-e5720312f674" containerName="heat-db-sync" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.055671 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1ae752-b331-4d0f-8039-e5720312f674" containerName="heat-db-sync" Dec 05 08:36:31 crc kubenswrapper[4876]: E1205 08:36:31.055689 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62f1d7bc-1a36-4fc0-a94d-72fcedda052e" containerName="cinder-db-sync" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.055698 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="62f1d7bc-1a36-4fc0-a94d-72fcedda052e" containerName="cinder-db-sync" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.056067 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="62f1d7bc-1a36-4fc0-a94d-72fcedda052e" containerName="cinder-db-sync" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.056086 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1ae752-b331-4d0f-8039-e5720312f674" containerName="heat-db-sync" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.058287 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.063668 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.067218 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.067500 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-hmxgp" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.067639 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.113365 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.207763 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-s2gmz"] Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.208011 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" podUID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerName="dnsmasq-dns" containerID="cri-o://19c25aedd431a59ef04a19e058aa9006fbc8142b46d2cb97b8db5944e08b6a96" gracePeriod=10 Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.211173 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.217288 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.217360 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxbxw\" (UniqueName: \"kubernetes.io/projected/d0f04750-4935-46c6-90ed-bf688288d04a-kube-api-access-cxbxw\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.217404 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.217444 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0f04750-4935-46c6-90ed-bf688288d04a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.217493 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.217543 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.271646 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2czdq"] Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.273398 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.285233 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2czdq"] Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.316979 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.319235 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.325581 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.327043 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.327086 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0f04750-4935-46c6-90ed-bf688288d04a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.327120 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.327157 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.327225 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.327257 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxbxw\" (UniqueName: \"kubernetes.io/projected/d0f04750-4935-46c6-90ed-bf688288d04a-kube-api-access-cxbxw\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.337419 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.337638 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0f04750-4935-46c6-90ed-bf688288d04a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.367967 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.368038 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.369138 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxbxw\" (UniqueName: \"kubernetes.io/projected/d0f04750-4935-46c6-90ed-bf688288d04a-kube-api-access-cxbxw\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.381536 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.399253 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.403287 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437397 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437444 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpqsx\" (UniqueName: \"kubernetes.io/projected/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-kube-api-access-rpqsx\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437473 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-config\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437499 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437519 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437563 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data-custom\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437583 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-scripts\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437608 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmgcf\" (UniqueName: \"kubernetes.io/projected/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-kube-api-access-nmgcf\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437627 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437649 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437676 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437692 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-logs\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.437709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.540818 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-config\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.540877 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.540916 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.540974 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data-custom\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.540995 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-scripts\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.541023 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmgcf\" (UniqueName: \"kubernetes.io/projected/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-kube-api-access-nmgcf\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.541046 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.541066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.541093 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.541110 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-logs\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.541128 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.541159 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.541180 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpqsx\" (UniqueName: \"kubernetes.io/projected/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-kube-api-access-rpqsx\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.542284 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-config\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.542768 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.545177 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.545465 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-logs\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.545558 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.546156 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.548556 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.549400 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.549781 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-scripts\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.553051 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" podUID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: connect: connection refused" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.553441 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.555423 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data-custom\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.578596 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmgcf\" (UniqueName: \"kubernetes.io/projected/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-kube-api-access-nmgcf\") pod \"dnsmasq-dns-6bb4fc677f-2czdq\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.581684 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpqsx\" (UniqueName: \"kubernetes.io/projected/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-kube-api-access-rpqsx\") pod \"cinder-api-0\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.618735 4876 generic.go:334] "Generic (PLEG): container finished" podID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerID="19c25aedd431a59ef04a19e058aa9006fbc8142b46d2cb97b8db5944e08b6a96" exitCode=0 Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.618837 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" event={"ID":"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c","Type":"ContainerDied","Data":"19c25aedd431a59ef04a19e058aa9006fbc8142b46d2cb97b8db5944e08b6a96"} Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.653675 4876 generic.go:334] "Generic (PLEG): container finished" podID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerID="a3451ab3ae1050cd70ba23c95cba113de1a07431dd1e3918f3f668bef54c4cd9" exitCode=0 Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.653702 4876 generic.go:334] "Generic (PLEG): container finished" podID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerID="f10cb06c492843f0aa303245e23e0d29c816b8c13819ee1f6780e7985654a987" exitCode=2 Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.653721 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5","Type":"ContainerDied","Data":"a3451ab3ae1050cd70ba23c95cba113de1a07431dd1e3918f3f668bef54c4cd9"} Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.653748 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5","Type":"ContainerDied","Data":"f10cb06c492843f0aa303245e23e0d29c816b8c13819ee1f6780e7985654a987"} Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.859303 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.883482 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:36:31 crc kubenswrapper[4876]: I1205 08:36:31.899674 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.061802 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-nb\") pod \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.061885 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-config\") pod \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.061945 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-sb\") pod \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.062099 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-swift-storage-0\") pod \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.062167 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-svc\") pod \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.062213 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4lz2\" (UniqueName: \"kubernetes.io/projected/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-kube-api-access-z4lz2\") pod \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\" (UID: \"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c\") " Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.079281 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-kube-api-access-z4lz2" (OuterVolumeSpecName: "kube-api-access-z4lz2") pod "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" (UID: "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c"). InnerVolumeSpecName "kube-api-access-z4lz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.138091 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.165165 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4lz2\" (UniqueName: \"kubernetes.io/projected/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-kube-api-access-z4lz2\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.182318 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" (UID: "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.186599 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" (UID: "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.196834 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" (UID: "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.197425 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" (UID: "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.212453 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-config" (OuterVolumeSpecName: "config") pod "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" (UID: "76c1d3f3-4815-4400-8b62-fff8d3bc0e3c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.269146 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.269462 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.269567 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.269670 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.269757 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:32 crc kubenswrapper[4876]: W1205 08:36:32.445771 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aca26c9_2dd2_46bb_ad3a_4b5e5bb4e07f.slice/crio-b00c4086cef3bbdd3bba6563f5785420ce514cdd89a0398e2d83f02f107e62ce WatchSource:0}: Error finding container b00c4086cef3bbdd3bba6563f5785420ce514cdd89a0398e2d83f02f107e62ce: Status 404 returned error can't find the container with id b00c4086cef3bbdd3bba6563f5785420ce514cdd89a0398e2d83f02f107e62ce Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.453806 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2czdq"] Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.633765 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.729997 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0f04750-4935-46c6-90ed-bf688288d04a","Type":"ContainerStarted","Data":"2db4da213233a2c2d7f7a3f2552be06e4bffc6c4c1382b83d7f2be656e31a3ac"} Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.752131 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd35f0e7-6d90-41a7-883d-6034ae6ead6d","Type":"ContainerStarted","Data":"b72990cf54151915a34e1b58b296d6eb1ea221f02da5b8e9c12508656f7e1d9c"} Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.762684 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" event={"ID":"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f","Type":"ContainerStarted","Data":"b00c4086cef3bbdd3bba6563f5785420ce514cdd89a0398e2d83f02f107e62ce"} Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.764493 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" event={"ID":"76c1d3f3-4815-4400-8b62-fff8d3bc0e3c","Type":"ContainerDied","Data":"269d8b3bc022898adcb8a965c1e0c0601da0d1d4b376b1a503e0677210857c81"} Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.764521 4876 scope.go:117] "RemoveContainer" containerID="19c25aedd431a59ef04a19e058aa9006fbc8142b46d2cb97b8db5944e08b6a96" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.764647 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-s2gmz" Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.807737 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-s2gmz"] Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.822886 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-s2gmz"] Dec 05 08:36:32 crc kubenswrapper[4876]: I1205 08:36:32.924923 4876 scope.go:117] "RemoveContainer" containerID="9374100375ae0cceb7ea0ab64339a9eabe2188c59fc65a757ab65efa8242303e" Dec 05 08:36:33 crc kubenswrapper[4876]: I1205 08:36:33.577301 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:33 crc kubenswrapper[4876]: I1205 08:36:33.645318 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:33 crc kubenswrapper[4876]: I1205 08:36:33.791273 4876 generic.go:334] "Generic (PLEG): container finished" podID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerID="bdb2749c0bdbf1d92e5bea8bf5cb1424c07fab5d7e2895e4bfcb0f9ab6f0c4d8" exitCode=0 Dec 05 08:36:33 crc kubenswrapper[4876]: I1205 08:36:33.791346 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5","Type":"ContainerDied","Data":"bdb2749c0bdbf1d92e5bea8bf5cb1424c07fab5d7e2895e4bfcb0f9ab6f0c4d8"} Dec 05 08:36:33 crc kubenswrapper[4876]: I1205 08:36:33.796643 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd35f0e7-6d90-41a7-883d-6034ae6ead6d","Type":"ContainerStarted","Data":"aa90966cd749eccbceda8f4302b54bfffdc1227a46f4dc80adca90f7d31650ce"} Dec 05 08:36:33 crc kubenswrapper[4876]: I1205 08:36:33.810424 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" event={"ID":"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f","Type":"ContainerStarted","Data":"ee8c99d68773c53b0b065da2715f54049a76de9a415e0b9d3eca801dd14691ca"} Dec 05 08:36:33 crc kubenswrapper[4876]: I1205 08:36:33.850015 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" path="/var/lib/kubelet/pods/76c1d3f3-4815-4400-8b62-fff8d3bc0e3c/volumes" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.146028 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.589037 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.669676 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-run-httpd\") pod \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.669760 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-config-data\") pod \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.669826 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-combined-ca-bundle\") pod \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.669854 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-scripts\") pod \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.669925 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5qgf\" (UniqueName: \"kubernetes.io/projected/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-kube-api-access-m5qgf\") pod \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.669959 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-log-httpd\") pod \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.670011 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-sg-core-conf-yaml\") pod \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\" (UID: \"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5\") " Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.673946 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" (UID: "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.674949 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" (UID: "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.676421 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-scripts" (OuterVolumeSpecName: "scripts") pod "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" (UID: "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.687005 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-kube-api-access-m5qgf" (OuterVolumeSpecName: "kube-api-access-m5qgf") pod "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" (UID: "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5"). InnerVolumeSpecName "kube-api-access-m5qgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.705513 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" (UID: "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.778398 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5qgf\" (UniqueName: \"kubernetes.io/projected/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-kube-api-access-m5qgf\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.778712 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.778722 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.778730 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.778738 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.783128 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-config-data" (OuterVolumeSpecName: "config-data") pod "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" (UID: "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.853393 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee3139f2-0c55-464e-bdbf-eefcc4bb59c5","Type":"ContainerDied","Data":"d8e3dbac459cdc2ae7796f6937bb3a55b8a27bf4ba24513b6215e35cad65a059"} Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.853448 4876 scope.go:117] "RemoveContainer" containerID="a3451ab3ae1050cd70ba23c95cba113de1a07431dd1e3918f3f668bef54c4cd9" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.853545 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.857140 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" (UID: "ee3139f2-0c55-464e-bdbf-eefcc4bb59c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.864419 4876 generic.go:334] "Generic (PLEG): container finished" podID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerID="ee8c99d68773c53b0b065da2715f54049a76de9a415e0b9d3eca801dd14691ca" exitCode=0 Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.864467 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" event={"ID":"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f","Type":"ContainerDied","Data":"ee8c99d68773c53b0b065da2715f54049a76de9a415e0b9d3eca801dd14691ca"} Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.880007 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:34 crc kubenswrapper[4876]: I1205 08:36:34.880038 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.105717 4876 scope.go:117] "RemoveContainer" containerID="f10cb06c492843f0aa303245e23e0d29c816b8c13819ee1f6780e7985654a987" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.156081 4876 scope.go:117] "RemoveContainer" containerID="bdb2749c0bdbf1d92e5bea8bf5cb1424c07fab5d7e2895e4bfcb0f9ab6f0c4d8" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.320025 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.336068 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.346011 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:36:35 crc kubenswrapper[4876]: E1205 08:36:35.346600 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="proxy-httpd" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.346696 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="proxy-httpd" Dec 05 08:36:35 crc kubenswrapper[4876]: E1205 08:36:35.346767 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="sg-core" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.346821 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="sg-core" Dec 05 08:36:35 crc kubenswrapper[4876]: E1205 08:36:35.346886 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerName="dnsmasq-dns" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.346980 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerName="dnsmasq-dns" Dec 05 08:36:35 crc kubenswrapper[4876]: E1205 08:36:35.347036 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="ceilometer-notification-agent" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.347089 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="ceilometer-notification-agent" Dec 05 08:36:35 crc kubenswrapper[4876]: E1205 08:36:35.347150 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerName="init" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.347204 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerName="init" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.347422 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="proxy-httpd" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.347492 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="sg-core" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.347550 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c1d3f3-4815-4400-8b62-fff8d3bc0e3c" containerName="dnsmasq-dns" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.347623 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" containerName="ceilometer-notification-agent" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.349352 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.352686 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.352991 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.368102 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.423815 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-run-httpd\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.424083 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.424165 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9tlh\" (UniqueName: \"kubernetes.io/projected/f00c9009-2533-4ced-8d54-aa866361aca1-kube-api-access-v9tlh\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.424279 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.424347 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-config-data\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.424414 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-scripts\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.424546 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-log-httpd\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.527823 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-log-httpd\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.527961 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-run-httpd\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.527989 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.528025 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9tlh\" (UniqueName: \"kubernetes.io/projected/f00c9009-2533-4ced-8d54-aa866361aca1-kube-api-access-v9tlh\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.528089 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.528108 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-config-data\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.528128 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-scripts\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.528891 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-run-httpd\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.529194 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-log-httpd\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.536592 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.537304 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.538523 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-config-data\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.550786 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9tlh\" (UniqueName: \"kubernetes.io/projected/f00c9009-2533-4ced-8d54-aa866361aca1-kube-api-access-v9tlh\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.552530 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-scripts\") pod \"ceilometer-0\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.672274 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:36:35 crc kubenswrapper[4876]: I1205 08:36:35.861041 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee3139f2-0c55-464e-bdbf-eefcc4bb59c5" path="/var/lib/kubelet/pods/ee3139f2-0c55-464e-bdbf-eefcc4bb59c5/volumes" Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.611101 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.691483 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:36:36 crc kubenswrapper[4876]: W1205 08:36:36.694547 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf00c9009_2533_4ced_8d54_aa866361aca1.slice/crio-726d7c8d8955e7c20391f73c73d55748ed3dffdcd12dcb24b52d1df4d6c0ff72 WatchSource:0}: Error finding container 726d7c8d8955e7c20391f73c73d55748ed3dffdcd12dcb24b52d1df4d6c0ff72: Status 404 returned error can't find the container with id 726d7c8d8955e7c20391f73c73d55748ed3dffdcd12dcb24b52d1df4d6c0ff72 Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.826752 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9d7c76978-sl6zr" Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.895725 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c5857d458-hs895"] Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.896008 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c5857d458-hs895" podUID="53d63901-e276-47e4-b544-5763ff670a20" containerName="barbican-api-log" containerID="cri-o://ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7" gracePeriod=30 Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.896847 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7c5857d458-hs895" podUID="53d63901-e276-47e4-b544-5763ff670a20" containerName="barbican-api" containerID="cri-o://735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d" gracePeriod=30 Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.935238 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerStarted","Data":"726d7c8d8955e7c20391f73c73d55748ed3dffdcd12dcb24b52d1df4d6c0ff72"} Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.937374 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" event={"ID":"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f","Type":"ContainerStarted","Data":"1da4a1b0422093f36912f7cf0a1894e8e02880a75acd79d72d15cb5fd2faa6d9"} Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.937510 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.952145 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0f04750-4935-46c6-90ed-bf688288d04a","Type":"ContainerStarted","Data":"df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e"} Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.963011 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd35f0e7-6d90-41a7-883d-6034ae6ead6d","Type":"ContainerStarted","Data":"d2d6df7e796a87501a0a54612f26f0b6405e95573cc70ae5db3a6844e97a5849"} Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.963141 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerName="cinder-api-log" containerID="cri-o://aa90966cd749eccbceda8f4302b54bfffdc1227a46f4dc80adca90f7d31650ce" gracePeriod=30 Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.963298 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerName="cinder-api" containerID="cri-o://d2d6df7e796a87501a0a54612f26f0b6405e95573cc70ae5db3a6844e97a5849" gracePeriod=30 Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.963451 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 08:36:36 crc kubenswrapper[4876]: I1205 08:36:36.977032 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" podStartSLOduration=5.977009998 podStartE2EDuration="5.977009998s" podCreationTimestamp="2025-12-05 08:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:36.95952663 +0000 UTC m=+1261.448191272" watchObservedRunningTime="2025-12-05 08:36:36.977009998 +0000 UTC m=+1261.465674630" Dec 05 08:36:37 crc kubenswrapper[4876]: I1205 08:36:37.037640 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.037617057 podStartE2EDuration="6.037617057s" podCreationTimestamp="2025-12-05 08:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:36.990981937 +0000 UTC m=+1261.479646559" watchObservedRunningTime="2025-12-05 08:36:37.037617057 +0000 UTC m=+1261.526281679" Dec 05 08:36:37 crc kubenswrapper[4876]: I1205 08:36:37.989064 4876 generic.go:334] "Generic (PLEG): container finished" podID="53d63901-e276-47e4-b544-5763ff670a20" containerID="ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7" exitCode=143 Dec 05 08:36:37 crc kubenswrapper[4876]: I1205 08:36:37.989398 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c5857d458-hs895" event={"ID":"53d63901-e276-47e4-b544-5763ff670a20","Type":"ContainerDied","Data":"ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7"} Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.004793 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0f04750-4935-46c6-90ed-bf688288d04a","Type":"ContainerStarted","Data":"6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a"} Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.026025 4876 generic.go:334] "Generic (PLEG): container finished" podID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerID="d2d6df7e796a87501a0a54612f26f0b6405e95573cc70ae5db3a6844e97a5849" exitCode=0 Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.026053 4876 generic.go:334] "Generic (PLEG): container finished" podID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerID="aa90966cd749eccbceda8f4302b54bfffdc1227a46f4dc80adca90f7d31650ce" exitCode=143 Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.026733 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd35f0e7-6d90-41a7-883d-6034ae6ead6d","Type":"ContainerDied","Data":"d2d6df7e796a87501a0a54612f26f0b6405e95573cc70ae5db3a6844e97a5849"} Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.026769 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd35f0e7-6d90-41a7-883d-6034ae6ead6d","Type":"ContainerDied","Data":"aa90966cd749eccbceda8f4302b54bfffdc1227a46f4dc80adca90f7d31650ce"} Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.033281 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.353358753 podStartE2EDuration="7.033270598s" podCreationTimestamp="2025-12-05 08:36:31 +0000 UTC" firstStartedPulling="2025-12-05 08:36:32.168423634 +0000 UTC m=+1256.657088256" lastFinishedPulling="2025-12-05 08:36:34.848335479 +0000 UTC m=+1259.337000101" observedRunningTime="2025-12-05 08:36:38.030452258 +0000 UTC m=+1262.519116880" watchObservedRunningTime="2025-12-05 08:36:38.033270598 +0000 UTC m=+1262.521935220" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.156605 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.213447 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.213500 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340252 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-etc-machine-id\") pod \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340381 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-logs\") pod \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340415 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-combined-ca-bundle\") pod \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340432 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data\") pod \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340456 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpqsx\" (UniqueName: \"kubernetes.io/projected/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-kube-api-access-rpqsx\") pod \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340496 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-scripts\") pod \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340522 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data-custom\") pod \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\" (UID: \"bd35f0e7-6d90-41a7-883d-6034ae6ead6d\") " Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340368 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bd35f0e7-6d90-41a7-883d-6034ae6ead6d" (UID: "bd35f0e7-6d90-41a7-883d-6034ae6ead6d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.340776 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-logs" (OuterVolumeSpecName: "logs") pod "bd35f0e7-6d90-41a7-883d-6034ae6ead6d" (UID: "bd35f0e7-6d90-41a7-883d-6034ae6ead6d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.341341 4876 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.341370 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.346608 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-scripts" (OuterVolumeSpecName: "scripts") pod "bd35f0e7-6d90-41a7-883d-6034ae6ead6d" (UID: "bd35f0e7-6d90-41a7-883d-6034ae6ead6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.346779 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bd35f0e7-6d90-41a7-883d-6034ae6ead6d" (UID: "bd35f0e7-6d90-41a7-883d-6034ae6ead6d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.349056 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-kube-api-access-rpqsx" (OuterVolumeSpecName: "kube-api-access-rpqsx") pod "bd35f0e7-6d90-41a7-883d-6034ae6ead6d" (UID: "bd35f0e7-6d90-41a7-883d-6034ae6ead6d"). InnerVolumeSpecName "kube-api-access-rpqsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.372992 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd35f0e7-6d90-41a7-883d-6034ae6ead6d" (UID: "bd35f0e7-6d90-41a7-883d-6034ae6ead6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.404111 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data" (OuterVolumeSpecName: "config-data") pod "bd35f0e7-6d90-41a7-883d-6034ae6ead6d" (UID: "bd35f0e7-6d90-41a7-883d-6034ae6ead6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.442524 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.442557 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.442568 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpqsx\" (UniqueName: \"kubernetes.io/projected/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-kube-api-access-rpqsx\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.442578 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:38 crc kubenswrapper[4876]: I1205 08:36:38.442586 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd35f0e7-6d90-41a7-883d-6034ae6ead6d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.038149 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerStarted","Data":"62ba626fcbf5cebe0178ed0f725dfe5c4b1921e502047ef0e1ec60a1d6e192ed"} Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.038459 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerStarted","Data":"9cf53475653d7b935071f3a802715b2a47167c1a284f665854d470eb11798148"} Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.040285 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"bd35f0e7-6d90-41a7-883d-6034ae6ead6d","Type":"ContainerDied","Data":"b72990cf54151915a34e1b58b296d6eb1ea221f02da5b8e9c12508656f7e1d9c"} Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.040307 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.040333 4876 scope.go:117] "RemoveContainer" containerID="d2d6df7e796a87501a0a54612f26f0b6405e95573cc70ae5db3a6844e97a5849" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.069536 4876 scope.go:117] "RemoveContainer" containerID="aa90966cd749eccbceda8f4302b54bfffdc1227a46f4dc80adca90f7d31650ce" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.090353 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.101154 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.144383 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:39 crc kubenswrapper[4876]: E1205 08:36:39.144881 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerName="cinder-api" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.144914 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerName="cinder-api" Dec 05 08:36:39 crc kubenswrapper[4876]: E1205 08:36:39.144949 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerName="cinder-api-log" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.144957 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerName="cinder-api-log" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.145181 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerName="cinder-api" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.145213 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" containerName="cinder-api-log" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.146349 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.152526 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.153263 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.155797 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.161295 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.258342 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e321464-50bf-42c0-8584-a75e9282eb65-logs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.258490 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.258624 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-config-data\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.258736 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.258844 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e321464-50bf-42c0-8584-a75e9282eb65-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.258959 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-config-data-custom\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.259150 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4xkf\" (UniqueName: \"kubernetes.io/projected/9e321464-50bf-42c0-8584-a75e9282eb65-kube-api-access-h4xkf\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.259431 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.259495 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-scripts\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.343080 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361152 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-config-data\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361275 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361344 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e321464-50bf-42c0-8584-a75e9282eb65-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361397 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-config-data-custom\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361445 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4xkf\" (UniqueName: \"kubernetes.io/projected/9e321464-50bf-42c0-8584-a75e9282eb65-kube-api-access-h4xkf\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361505 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e321464-50bf-42c0-8584-a75e9282eb65-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361564 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361616 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-scripts\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361676 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e321464-50bf-42c0-8584-a75e9282eb65-logs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.361712 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.362192 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e321464-50bf-42c0-8584-a75e9282eb65-logs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.366010 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.366265 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-scripts\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.367159 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.368654 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-config-data-custom\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.369482 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.397660 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e321464-50bf-42c0-8584-a75e9282eb65-config-data\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.404542 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4xkf\" (UniqueName: \"kubernetes.io/projected/9e321464-50bf-42c0-8584-a75e9282eb65-kube-api-access-h4xkf\") pod \"cinder-api-0\" (UID: \"9e321464-50bf-42c0-8584-a75e9282eb65\") " pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.481048 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.836817 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd35f0e7-6d90-41a7-883d-6034ae6ead6d" path="/var/lib/kubelet/pods/bd35f0e7-6d90-41a7-883d-6034ae6ead6d/volumes" Dec 05 08:36:39 crc kubenswrapper[4876]: I1205 08:36:39.966635 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.051118 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9e321464-50bf-42c0-8584-a75e9282eb65","Type":"ContainerStarted","Data":"df9e7fdd10a2f6dbfffa00ca975ee5f37ddfa12cdf81010c6874f7d0d0a403ec"} Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.059796 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerStarted","Data":"caf8ef8141c08c3c5eae3d32464d7773ae19090c1360c96e0e03ede73137bb41"} Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.496691 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.692533 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxcdh\" (UniqueName: \"kubernetes.io/projected/53d63901-e276-47e4-b544-5763ff670a20-kube-api-access-jxcdh\") pod \"53d63901-e276-47e4-b544-5763ff670a20\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.692645 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data-custom\") pod \"53d63901-e276-47e4-b544-5763ff670a20\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.692692 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data\") pod \"53d63901-e276-47e4-b544-5763ff670a20\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.692796 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d63901-e276-47e4-b544-5763ff670a20-logs\") pod \"53d63901-e276-47e4-b544-5763ff670a20\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.692998 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-combined-ca-bundle\") pod \"53d63901-e276-47e4-b544-5763ff670a20\" (UID: \"53d63901-e276-47e4-b544-5763ff670a20\") " Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.697298 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53d63901-e276-47e4-b544-5763ff670a20-logs" (OuterVolumeSpecName: "logs") pod "53d63901-e276-47e4-b544-5763ff670a20" (UID: "53d63901-e276-47e4-b544-5763ff670a20"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.706059 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "53d63901-e276-47e4-b544-5763ff670a20" (UID: "53d63901-e276-47e4-b544-5763ff670a20"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.744210 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d63901-e276-47e4-b544-5763ff670a20-kube-api-access-jxcdh" (OuterVolumeSpecName: "kube-api-access-jxcdh") pod "53d63901-e276-47e4-b544-5763ff670a20" (UID: "53d63901-e276-47e4-b544-5763ff670a20"). InnerVolumeSpecName "kube-api-access-jxcdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.796991 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d63901-e276-47e4-b544-5763ff670a20-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.797033 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxcdh\" (UniqueName: \"kubernetes.io/projected/53d63901-e276-47e4-b544-5763ff670a20-kube-api-access-jxcdh\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.797045 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.834735 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53d63901-e276-47e4-b544-5763ff670a20" (UID: "53d63901-e276-47e4-b544-5763ff670a20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.858630 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data" (OuterVolumeSpecName: "config-data") pod "53d63901-e276-47e4-b544-5763ff670a20" (UID: "53d63901-e276-47e4-b544-5763ff670a20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.900174 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:40 crc kubenswrapper[4876]: I1205 08:36:40.900530 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d63901-e276-47e4-b544-5763ff670a20-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.072950 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9e321464-50bf-42c0-8584-a75e9282eb65","Type":"ContainerStarted","Data":"689d27278e6de493eb99660f1f384bb944a67ddc5075dc08e5322c00a95fad7e"} Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.076605 4876 generic.go:334] "Generic (PLEG): container finished" podID="53d63901-e276-47e4-b544-5763ff670a20" containerID="735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d" exitCode=0 Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.076639 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c5857d458-hs895" event={"ID":"53d63901-e276-47e4-b544-5763ff670a20","Type":"ContainerDied","Data":"735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d"} Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.076656 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7c5857d458-hs895" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.076666 4876 scope.go:117] "RemoveContainer" containerID="735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.076657 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7c5857d458-hs895" event={"ID":"53d63901-e276-47e4-b544-5763ff670a20","Type":"ContainerDied","Data":"380ec081a95e02b3a178730d9e5691f2940c131360e1e0c7a01b5542cc34dbc9"} Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.104159 4876 scope.go:117] "RemoveContainer" containerID="ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.131102 4876 scope.go:117] "RemoveContainer" containerID="735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d" Dec 05 08:36:41 crc kubenswrapper[4876]: E1205 08:36:41.132353 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d\": container with ID starting with 735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d not found: ID does not exist" containerID="735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.132391 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d"} err="failed to get container status \"735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d\": rpc error: code = NotFound desc = could not find container \"735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d\": container with ID starting with 735d7b37b104323626ce2a556222ed22830f196a53340a3687f7a0db3fe4fa4d not found: ID does not exist" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.132415 4876 scope.go:117] "RemoveContainer" containerID="ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7" Dec 05 08:36:41 crc kubenswrapper[4876]: E1205 08:36:41.132642 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7\": container with ID starting with ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7 not found: ID does not exist" containerID="ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.132665 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7"} err="failed to get container status \"ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7\": rpc error: code = NotFound desc = could not find container \"ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7\": container with ID starting with ade188d81764cd4450d015995b3982ab6a20874c379dab669e63b4fcba8812e7 not found: ID does not exist" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.137663 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7c5857d458-hs895"] Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.147174 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7c5857d458-hs895"] Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.404261 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.616315 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.835275 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d63901-e276-47e4-b544-5763ff670a20" path="/var/lib/kubelet/pods/53d63901-e276-47e4-b544-5763ff670a20/volumes" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.862053 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.935240 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-4qckc"] Dec 05 08:36:41 crc kubenswrapper[4876]: I1205 08:36:41.935530 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" podUID="7eb4b169-060a-45b0-8607-7538547b5640" containerName="dnsmasq-dns" containerID="cri-o://35822f26df723a06788493800f3466d8a9149a0df877825b531946c6eb9868d6" gracePeriod=10 Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.096149 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerStarted","Data":"30eb4d1c6734bfaaf784d91bc6fe47ae09afeb912947fcf232f72666387342d2"} Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.097347 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.105948 4876 generic.go:334] "Generic (PLEG): container finished" podID="7eb4b169-060a-45b0-8607-7538547b5640" containerID="35822f26df723a06788493800f3466d8a9149a0df877825b531946c6eb9868d6" exitCode=0 Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.106310 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" event={"ID":"7eb4b169-060a-45b0-8607-7538547b5640","Type":"ContainerDied","Data":"35822f26df723a06788493800f3466d8a9149a0df877825b531946c6eb9868d6"} Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.110066 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9e321464-50bf-42c0-8584-a75e9282eb65","Type":"ContainerStarted","Data":"04f19b231894d978596734baa29d0011de0521c5b4110bed5e90cfcf06b51cfc"} Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.110189 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.110327 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5ff8f8c8c7-rx6w7" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.130850 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.937135987 podStartE2EDuration="7.130830821s" podCreationTimestamp="2025-12-05 08:36:35 +0000 UTC" firstStartedPulling="2025-12-05 08:36:36.696808166 +0000 UTC m=+1261.185472788" lastFinishedPulling="2025-12-05 08:36:40.890503 +0000 UTC m=+1265.379167622" observedRunningTime="2025-12-05 08:36:42.127749683 +0000 UTC m=+1266.616414295" watchObservedRunningTime="2025-12-05 08:36:42.130830821 +0000 UTC m=+1266.619495443" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.176812 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-75d658964-r6f2m"] Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.177050 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-75d658964-r6f2m" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerName="neutron-api" containerID="cri-o://65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362" gracePeriod=30 Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.177191 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-75d658964-r6f2m" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerName="neutron-httpd" containerID="cri-o://6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc" gracePeriod=30 Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.216219 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.231934 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.231914994 podStartE2EDuration="3.231914994s" podCreationTimestamp="2025-12-05 08:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:42.216373901 +0000 UTC m=+1266.705038523" watchObservedRunningTime="2025-12-05 08:36:42.231914994 +0000 UTC m=+1266.720579616" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.563042 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.637988 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-config\") pod \"7eb4b169-060a-45b0-8607-7538547b5640\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.638088 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-svc\") pod \"7eb4b169-060a-45b0-8607-7538547b5640\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.638120 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rctk6\" (UniqueName: \"kubernetes.io/projected/7eb4b169-060a-45b0-8607-7538547b5640-kube-api-access-rctk6\") pod \"7eb4b169-060a-45b0-8607-7538547b5640\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.638217 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-nb\") pod \"7eb4b169-060a-45b0-8607-7538547b5640\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.638252 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-sb\") pod \"7eb4b169-060a-45b0-8607-7538547b5640\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.638287 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-swift-storage-0\") pod \"7eb4b169-060a-45b0-8607-7538547b5640\" (UID: \"7eb4b169-060a-45b0-8607-7538547b5640\") " Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.659182 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb4b169-060a-45b0-8607-7538547b5640-kube-api-access-rctk6" (OuterVolumeSpecName: "kube-api-access-rctk6") pod "7eb4b169-060a-45b0-8607-7538547b5640" (UID: "7eb4b169-060a-45b0-8607-7538547b5640"). InnerVolumeSpecName "kube-api-access-rctk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.703974 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7eb4b169-060a-45b0-8607-7538547b5640" (UID: "7eb4b169-060a-45b0-8607-7538547b5640"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.708516 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7eb4b169-060a-45b0-8607-7538547b5640" (UID: "7eb4b169-060a-45b0-8607-7538547b5640"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.715530 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7eb4b169-060a-45b0-8607-7538547b5640" (UID: "7eb4b169-060a-45b0-8607-7538547b5640"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.719570 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7eb4b169-060a-45b0-8607-7538547b5640" (UID: "7eb4b169-060a-45b0-8607-7538547b5640"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.734597 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-config" (OuterVolumeSpecName: "config") pod "7eb4b169-060a-45b0-8607-7538547b5640" (UID: "7eb4b169-060a-45b0-8607-7538547b5640"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.740190 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.740217 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rctk6\" (UniqueName: \"kubernetes.io/projected/7eb4b169-060a-45b0-8607-7538547b5640-kube-api-access-rctk6\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.740228 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.740237 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.740355 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:42 crc kubenswrapper[4876]: I1205 08:36:42.740366 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7eb4b169-060a-45b0-8607-7538547b5640-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.124933 4876 generic.go:334] "Generic (PLEG): container finished" podID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerID="6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc" exitCode=0 Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.124983 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75d658964-r6f2m" event={"ID":"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0","Type":"ContainerDied","Data":"6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc"} Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.127277 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" event={"ID":"7eb4b169-060a-45b0-8607-7538547b5640","Type":"ContainerDied","Data":"08bc39ed30a8c882534c885d74a0013415387dc83b5acd6b5e7209727cdd6e79"} Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.127324 4876 scope.go:117] "RemoveContainer" containerID="35822f26df723a06788493800f3466d8a9149a0df877825b531946c6eb9868d6" Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.127467 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" containerName="cinder-scheduler" containerID="cri-o://df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e" gracePeriod=30 Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.127506 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-4qckc" Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.129421 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" containerName="probe" containerID="cri-o://6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a" gracePeriod=30 Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.157564 4876 scope.go:117] "RemoveContainer" containerID="5c0257f1ae3521b305faf224679b6a31da84614003320d50a3aef367435f9c7c" Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.179746 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-4qckc"] Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.188804 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-4qckc"] Dec 05 08:36:43 crc kubenswrapper[4876]: I1205 08:36:43.842168 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eb4b169-060a-45b0-8607-7538547b5640" path="/var/lib/kubelet/pods/7eb4b169-060a-45b0-8607-7538547b5640/volumes" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.138306 4876 generic.go:334] "Generic (PLEG): container finished" podID="d0f04750-4935-46c6-90ed-bf688288d04a" containerID="df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e" exitCode=0 Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.138435 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0f04750-4935-46c6-90ed-bf688288d04a","Type":"ContainerDied","Data":"df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e"} Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.570734 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.678068 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-combined-ca-bundle\") pod \"d0f04750-4935-46c6-90ed-bf688288d04a\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.678109 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data\") pod \"d0f04750-4935-46c6-90ed-bf688288d04a\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.678130 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data-custom\") pod \"d0f04750-4935-46c6-90ed-bf688288d04a\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.678164 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-scripts\") pod \"d0f04750-4935-46c6-90ed-bf688288d04a\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.678278 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxbxw\" (UniqueName: \"kubernetes.io/projected/d0f04750-4935-46c6-90ed-bf688288d04a-kube-api-access-cxbxw\") pod \"d0f04750-4935-46c6-90ed-bf688288d04a\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.678319 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0f04750-4935-46c6-90ed-bf688288d04a-etc-machine-id\") pod \"d0f04750-4935-46c6-90ed-bf688288d04a\" (UID: \"d0f04750-4935-46c6-90ed-bf688288d04a\") " Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.678447 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0f04750-4935-46c6-90ed-bf688288d04a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d0f04750-4935-46c6-90ed-bf688288d04a" (UID: "d0f04750-4935-46c6-90ed-bf688288d04a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.678679 4876 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0f04750-4935-46c6-90ed-bf688288d04a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.683969 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0f04750-4935-46c6-90ed-bf688288d04a-kube-api-access-cxbxw" (OuterVolumeSpecName: "kube-api-access-cxbxw") pod "d0f04750-4935-46c6-90ed-bf688288d04a" (UID: "d0f04750-4935-46c6-90ed-bf688288d04a"). InnerVolumeSpecName "kube-api-access-cxbxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.686008 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-scripts" (OuterVolumeSpecName: "scripts") pod "d0f04750-4935-46c6-90ed-bf688288d04a" (UID: "d0f04750-4935-46c6-90ed-bf688288d04a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.690470 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d0f04750-4935-46c6-90ed-bf688288d04a" (UID: "d0f04750-4935-46c6-90ed-bf688288d04a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.739547 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0f04750-4935-46c6-90ed-bf688288d04a" (UID: "d0f04750-4935-46c6-90ed-bf688288d04a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.781068 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.781105 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.781116 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.781126 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxbxw\" (UniqueName: \"kubernetes.io/projected/d0f04750-4935-46c6-90ed-bf688288d04a-kube-api-access-cxbxw\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.810140 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data" (OuterVolumeSpecName: "config-data") pod "d0f04750-4935-46c6-90ed-bf688288d04a" (UID: "d0f04750-4935-46c6-90ed-bf688288d04a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.882999 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0f04750-4935-46c6-90ed-bf688288d04a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:44 crc kubenswrapper[4876]: I1205 08:36:44.936574 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.086090 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-httpd-config\") pod \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.086244 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9chf\" (UniqueName: \"kubernetes.io/projected/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-kube-api-access-c9chf\") pod \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.086313 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-ovndb-tls-certs\") pod \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.086356 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-config\") pod \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.086477 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-combined-ca-bundle\") pod \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\" (UID: \"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0\") " Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.089748 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" (UID: "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.090971 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-kube-api-access-c9chf" (OuterVolumeSpecName: "kube-api-access-c9chf") pod "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" (UID: "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0"). InnerVolumeSpecName "kube-api-access-c9chf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.138159 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-config" (OuterVolumeSpecName: "config") pod "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" (UID: "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.154523 4876 generic.go:334] "Generic (PLEG): container finished" podID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerID="65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362" exitCode=0 Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.154583 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75d658964-r6f2m" event={"ID":"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0","Type":"ContainerDied","Data":"65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362"} Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.154670 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75d658964-r6f2m" event={"ID":"f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0","Type":"ContainerDied","Data":"64fc9f8a6aae6fb1d89781e4d3af3b455d6fb467d5edb3ce5fae1cae2011f061"} Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.154582 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75d658964-r6f2m" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.154927 4876 scope.go:117] "RemoveContainer" containerID="6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.165025 4876 generic.go:334] "Generic (PLEG): container finished" podID="d0f04750-4935-46c6-90ed-bf688288d04a" containerID="6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a" exitCode=0 Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.165073 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0f04750-4935-46c6-90ed-bf688288d04a","Type":"ContainerDied","Data":"6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a"} Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.165109 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0f04750-4935-46c6-90ed-bf688288d04a","Type":"ContainerDied","Data":"2db4da213233a2c2d7f7a3f2552be06e4bffc6c4c1382b83d7f2be656e31a3ac"} Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.165106 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.171314 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" (UID: "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.189937 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.189978 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.189994 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9chf\" (UniqueName: \"kubernetes.io/projected/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-kube-api-access-c9chf\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.190007 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.191230 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" (UID: "f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.292207 4876 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.325823 4876 scope.go:117] "RemoveContainer" containerID="65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.336373 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.356547 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.367624 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.368005 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" containerName="probe" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368022 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" containerName="probe" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.368032 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb4b169-060a-45b0-8607-7538547b5640" containerName="init" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368038 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb4b169-060a-45b0-8607-7538547b5640" containerName="init" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.368050 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d63901-e276-47e4-b544-5763ff670a20" containerName="barbican-api-log" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368056 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d63901-e276-47e4-b544-5763ff670a20" containerName="barbican-api-log" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.368074 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerName="neutron-httpd" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368080 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerName="neutron-httpd" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.368097 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d63901-e276-47e4-b544-5763ff670a20" containerName="barbican-api" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368103 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d63901-e276-47e4-b544-5763ff670a20" containerName="barbican-api" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.368118 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb4b169-060a-45b0-8607-7538547b5640" containerName="dnsmasq-dns" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368124 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb4b169-060a-45b0-8607-7538547b5640" containerName="dnsmasq-dns" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.368136 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" containerName="cinder-scheduler" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368142 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" containerName="cinder-scheduler" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.368152 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerName="neutron-api" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368157 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerName="neutron-api" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368303 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d63901-e276-47e4-b544-5763ff670a20" containerName="barbican-api" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368311 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" containerName="probe" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368320 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" containerName="cinder-scheduler" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368356 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eb4b169-060a-45b0-8607-7538547b5640" containerName="dnsmasq-dns" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368369 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerName="neutron-httpd" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368380 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" containerName="neutron-api" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.368392 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d63901-e276-47e4-b544-5763ff670a20" containerName="barbican-api-log" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.370161 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.375985 4876 scope.go:117] "RemoveContainer" containerID="6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.378030 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.381277 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc\": container with ID starting with 6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc not found: ID does not exist" containerID="6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.381317 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc"} err="failed to get container status \"6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc\": rpc error: code = NotFound desc = could not find container \"6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc\": container with ID starting with 6fbe5e7aa01649f7e729cdbc7217f69957f609e233fc16a75d578c5ba24a21fc not found: ID does not exist" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.381345 4876 scope.go:117] "RemoveContainer" containerID="65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.381814 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362\": container with ID starting with 65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362 not found: ID does not exist" containerID="65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.381872 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362"} err="failed to get container status \"65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362\": rpc error: code = NotFound desc = could not find container \"65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362\": container with ID starting with 65252fb37e4a6f9e2d1f56d0e599710b770b663bc826b2c87a7f81e3b9ad1362 not found: ID does not exist" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.381914 4876 scope.go:117] "RemoveContainer" containerID="6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.398725 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.402628 4876 scope.go:117] "RemoveContainer" containerID="df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.431548 4876 scope.go:117] "RemoveContainer" containerID="6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.434782 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a\": container with ID starting with 6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a not found: ID does not exist" containerID="6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.434826 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a"} err="failed to get container status \"6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a\": rpc error: code = NotFound desc = could not find container \"6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a\": container with ID starting with 6437fe363ebcec46ef645264e2b9fb8f7f6fe370528643c4ae164ca853752a7a not found: ID does not exist" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.434858 4876 scope.go:117] "RemoveContainer" containerID="df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e" Dec 05 08:36:45 crc kubenswrapper[4876]: E1205 08:36:45.435411 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e\": container with ID starting with df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e not found: ID does not exist" containerID="df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.435443 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e"} err="failed to get container status \"df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e\": rpc error: code = NotFound desc = could not find container \"df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e\": container with ID starting with df74a7b66c930a25e64ee5059cf1b970c68392a1fffaed64f33c131abbf9530e not found: ID does not exist" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.488803 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-75d658964-r6f2m"] Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.496044 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-75d658964-r6f2m"] Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.496172 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-config-data\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.496209 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.496227 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba584dcd-132e-407b-960e-4a68cfb1ea31-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.496287 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pkxp\" (UniqueName: \"kubernetes.io/projected/ba584dcd-132e-407b-960e-4a68cfb1ea31-kube-api-access-5pkxp\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.496309 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.496329 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-scripts\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.597593 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-config-data\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.597660 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.597685 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba584dcd-132e-407b-960e-4a68cfb1ea31-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.597776 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pkxp\" (UniqueName: \"kubernetes.io/projected/ba584dcd-132e-407b-960e-4a68cfb1ea31-kube-api-access-5pkxp\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.597807 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.597833 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-scripts\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.598148 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba584dcd-132e-407b-960e-4a68cfb1ea31-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.609808 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-config-data\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.610330 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-scripts\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.610330 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.612135 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba584dcd-132e-407b-960e-4a68cfb1ea31-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.613043 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pkxp\" (UniqueName: \"kubernetes.io/projected/ba584dcd-132e-407b-960e-4a68cfb1ea31-kube-api-access-5pkxp\") pod \"cinder-scheduler-0\" (UID: \"ba584dcd-132e-407b-960e-4a68cfb1ea31\") " pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.691883 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.858766 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0f04750-4935-46c6-90ed-bf688288d04a" path="/var/lib/kubelet/pods/d0f04750-4935-46c6-90ed-bf688288d04a/volumes" Dec 05 08:36:45 crc kubenswrapper[4876]: I1205 08:36:45.859556 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0" path="/var/lib/kubelet/pods/f6b7c7d0-1dd2-4fb6-9e9d-282f707528c0/volumes" Dec 05 08:36:46 crc kubenswrapper[4876]: I1205 08:36:46.172730 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 08:36:46 crc kubenswrapper[4876]: W1205 08:36:46.181325 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba584dcd_132e_407b_960e_4a68cfb1ea31.slice/crio-63bea75672e9d931e2cf59b3aebda8f757e3d172b311d1df594a530b9cb084fa WatchSource:0}: Error finding container 63bea75672e9d931e2cf59b3aebda8f757e3d172b311d1df594a530b9cb084fa: Status 404 returned error can't find the container with id 63bea75672e9d931e2cf59b3aebda8f757e3d172b311d1df594a530b9cb084fa Dec 05 08:36:47 crc kubenswrapper[4876]: I1205 08:36:47.194044 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ba584dcd-132e-407b-960e-4a68cfb1ea31","Type":"ContainerStarted","Data":"04466cf64c6db016b52cf3b61bfc7045a07ab4378eb2c59841431baa9234ff87"} Dec 05 08:36:47 crc kubenswrapper[4876]: I1205 08:36:47.194340 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ba584dcd-132e-407b-960e-4a68cfb1ea31","Type":"ContainerStarted","Data":"63bea75672e9d931e2cf59b3aebda8f757e3d172b311d1df594a530b9cb084fa"} Dec 05 08:36:48 crc kubenswrapper[4876]: I1205 08:36:48.203775 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ba584dcd-132e-407b-960e-4a68cfb1ea31","Type":"ContainerStarted","Data":"96627877cc7fe69112d94761a8e0aab750ab6c197e33dbf2bde5a1d074e6b778"} Dec 05 08:36:48 crc kubenswrapper[4876]: I1205 08:36:48.237552 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.237532943 podStartE2EDuration="3.237532943s" podCreationTimestamp="2025-12-05 08:36:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:48.223707779 +0000 UTC m=+1272.712372471" watchObservedRunningTime="2025-12-05 08:36:48.237532943 +0000 UTC m=+1272.726197565" Dec 05 08:36:49 crc kubenswrapper[4876]: I1205 08:36:49.944983 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-577c476758-ksnrz" Dec 05 08:36:50 crc kubenswrapper[4876]: I1205 08:36:50.692886 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 08:36:51 crc kubenswrapper[4876]: I1205 08:36:51.745648 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.109413 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.110490 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.112543 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.116681 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.120557 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-nlz9d" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.130027 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.271045 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-combined-ca-bundle\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.271386 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sdxt\" (UniqueName: \"kubernetes.io/projected/86ac21be-c7c2-4035-a32c-3f24715c3248-kube-api-access-2sdxt\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.271476 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.271538 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config-secret\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.373001 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sdxt\" (UniqueName: \"kubernetes.io/projected/86ac21be-c7c2-4035-a32c-3f24715c3248-kube-api-access-2sdxt\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.373148 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.373228 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config-secret\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.373265 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-combined-ca-bundle\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.374868 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.381250 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-combined-ca-bundle\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.387334 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config-secret\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.403686 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sdxt\" (UniqueName: \"kubernetes.io/projected/86ac21be-c7c2-4035-a32c-3f24715c3248-kube-api-access-2sdxt\") pod \"openstackclient\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.428986 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.599820 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.651795 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-56d57865bd-mfljr" Dec 05 08:36:52 crc kubenswrapper[4876]: I1205 08:36:52.997676 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 08:36:53 crc kubenswrapper[4876]: I1205 08:36:53.262116 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"86ac21be-c7c2-4035-a32c-3f24715c3248","Type":"ContainerStarted","Data":"c0f9f658c2116d713c46be8e16506c018d9adda105fa4de71d733e8b1aadba2f"} Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.248511 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-689f7d9f48-fjhvt"] Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.249652 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.252205 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.253258 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-52ct9" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.253424 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.276364 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-689f7d9f48-fjhvt"] Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.344685 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.344828 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp5z9\" (UniqueName: \"kubernetes.io/projected/8b5c7095-36f0-4024-b430-714d5603e08b-kube-api-access-kp5z9\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.344886 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-combined-ca-bundle\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.344939 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data-custom\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.361498 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-tw8hl"] Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.381854 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.410468 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-tw8hl"] Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.445912 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-876689776-4sqds"] Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.446995 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-svc\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447036 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxvmt\" (UniqueName: \"kubernetes.io/projected/bd36d5f7-6681-4eba-a08c-28357a52cf5e-kube-api-access-fxvmt\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447053 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447075 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447115 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447179 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447214 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-config\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447237 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp5z9\" (UniqueName: \"kubernetes.io/projected/8b5c7095-36f0-4024-b430-714d5603e08b-kube-api-access-kp5z9\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447258 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-combined-ca-bundle\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.447276 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data-custom\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.449152 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.462648 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.462735 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data-custom\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.462771 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.467791 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-combined-ca-bundle\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.494242 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-876689776-4sqds"] Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.548627 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.548696 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-config\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.549709 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.550641 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-config\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.550765 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-combined-ca-bundle\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.551149 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-svc\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.551319 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cvjh\" (UniqueName: \"kubernetes.io/projected/579ebb91-13da-40a1-80f2-c245cb440580-kube-api-access-5cvjh\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.551425 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxvmt\" (UniqueName: \"kubernetes.io/projected/bd36d5f7-6681-4eba-a08c-28357a52cf5e-kube-api-access-fxvmt\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.551534 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.551636 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.551931 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.552064 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data-custom\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.552879 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-svc\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.553953 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.553960 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5846cfbdbf-5ctlt"] Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.554886 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.555625 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.563140 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5846cfbdbf-5ctlt"] Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.569297 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.571171 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp5z9\" (UniqueName: \"kubernetes.io/projected/8b5c7095-36f0-4024-b430-714d5603e08b-kube-api-access-kp5z9\") pod \"heat-engine-689f7d9f48-fjhvt\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.580890 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxvmt\" (UniqueName: \"kubernetes.io/projected/bd36d5f7-6681-4eba-a08c-28357a52cf5e-kube-api-access-fxvmt\") pod \"dnsmasq-dns-7d978555f9-tw8hl\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.583315 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.654860 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvpd2\" (UniqueName: \"kubernetes.io/projected/718f1c24-daf5-44f1-91f3-51890c190f47-kube-api-access-vvpd2\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.655271 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cvjh\" (UniqueName: \"kubernetes.io/projected/579ebb91-13da-40a1-80f2-c245cb440580-kube-api-access-5cvjh\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.655303 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.655329 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-combined-ca-bundle\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.655385 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.655413 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data-custom\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.655488 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-combined-ca-bundle\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.655509 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data-custom\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.673741 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.674296 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cvjh\" (UniqueName: \"kubernetes.io/projected/579ebb91-13da-40a1-80f2-c245cb440580-kube-api-access-5cvjh\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.682961 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data-custom\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.683595 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-combined-ca-bundle\") pod \"heat-cfnapi-876689776-4sqds\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.779579 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.780351 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data-custom\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.780538 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvpd2\" (UniqueName: \"kubernetes.io/projected/718f1c24-daf5-44f1-91f3-51890c190f47-kube-api-access-vvpd2\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.780603 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.780649 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-combined-ca-bundle\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.796437 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-combined-ca-bundle\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.802777 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.807655 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvpd2\" (UniqueName: \"kubernetes.io/projected/718f1c24-daf5-44f1-91f3-51890c190f47-kube-api-access-vvpd2\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.814173 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data-custom\") pod \"heat-api-5846cfbdbf-5ctlt\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:55 crc kubenswrapper[4876]: I1205 08:36:55.936637 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:36:56 crc kubenswrapper[4876]: I1205 08:36:56.081440 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:36:56 crc kubenswrapper[4876]: I1205 08:36:56.230989 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-689f7d9f48-fjhvt"] Dec 05 08:36:56 crc kubenswrapper[4876]: W1205 08:36:56.235701 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b5c7095_36f0_4024_b430_714d5603e08b.slice/crio-fbb76d64ed72e6febfa177079798497bfecf76631dd7b9eb3c8c6767bc6a5de3 WatchSource:0}: Error finding container fbb76d64ed72e6febfa177079798497bfecf76631dd7b9eb3c8c6767bc6a5de3: Status 404 returned error can't find the container with id fbb76d64ed72e6febfa177079798497bfecf76631dd7b9eb3c8c6767bc6a5de3 Dec 05 08:36:56 crc kubenswrapper[4876]: I1205 08:36:56.319534 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-689f7d9f48-fjhvt" event={"ID":"8b5c7095-36f0-4024-b430-714d5603e08b","Type":"ContainerStarted","Data":"fbb76d64ed72e6febfa177079798497bfecf76631dd7b9eb3c8c6767bc6a5de3"} Dec 05 08:36:56 crc kubenswrapper[4876]: I1205 08:36:56.349351 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 08:36:56 crc kubenswrapper[4876]: I1205 08:36:56.408587 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-tw8hl"] Dec 05 08:36:56 crc kubenswrapper[4876]: W1205 08:36:56.441272 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd36d5f7_6681_4eba_a08c_28357a52cf5e.slice/crio-028cbac9b6ca7009a4811604d5e7f4dcfdad322e0f64810764422fef7110ccd4 WatchSource:0}: Error finding container 028cbac9b6ca7009a4811604d5e7f4dcfdad322e0f64810764422fef7110ccd4: Status 404 returned error can't find the container with id 028cbac9b6ca7009a4811604d5e7f4dcfdad322e0f64810764422fef7110ccd4 Dec 05 08:36:56 crc kubenswrapper[4876]: I1205 08:36:56.650635 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-876689776-4sqds"] Dec 05 08:36:56 crc kubenswrapper[4876]: W1205 08:36:56.701829 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod579ebb91_13da_40a1_80f2_c245cb440580.slice/crio-d8bc27eb15aad0f3834ac0737a13d25d90d5a2dccfbbac4b9f7481199db98c6a WatchSource:0}: Error finding container d8bc27eb15aad0f3834ac0737a13d25d90d5a2dccfbbac4b9f7481199db98c6a: Status 404 returned error can't find the container with id d8bc27eb15aad0f3834ac0737a13d25d90d5a2dccfbbac4b9f7481199db98c6a Dec 05 08:36:56 crc kubenswrapper[4876]: I1205 08:36:56.770035 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5846cfbdbf-5ctlt"] Dec 05 08:36:56 crc kubenswrapper[4876]: W1205 08:36:56.779108 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718f1c24_daf5_44f1_91f3_51890c190f47.slice/crio-a0dea39fcfd63a0ee62580f31d0b71f1c035a8a072b989dbf7c79ab97ba59d0c WatchSource:0}: Error finding container a0dea39fcfd63a0ee62580f31d0b71f1c035a8a072b989dbf7c79ab97ba59d0c: Status 404 returned error can't find the container with id a0dea39fcfd63a0ee62580f31d0b71f1c035a8a072b989dbf7c79ab97ba59d0c Dec 05 08:36:57 crc kubenswrapper[4876]: I1205 08:36:57.334695 4876 generic.go:334] "Generic (PLEG): container finished" podID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerID="e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3" exitCode=0 Dec 05 08:36:57 crc kubenswrapper[4876]: I1205 08:36:57.334761 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" event={"ID":"bd36d5f7-6681-4eba-a08c-28357a52cf5e","Type":"ContainerDied","Data":"e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3"} Dec 05 08:36:57 crc kubenswrapper[4876]: I1205 08:36:57.334786 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" event={"ID":"bd36d5f7-6681-4eba-a08c-28357a52cf5e","Type":"ContainerStarted","Data":"028cbac9b6ca7009a4811604d5e7f4dcfdad322e0f64810764422fef7110ccd4"} Dec 05 08:36:57 crc kubenswrapper[4876]: I1205 08:36:57.342088 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5846cfbdbf-5ctlt" event={"ID":"718f1c24-daf5-44f1-91f3-51890c190f47","Type":"ContainerStarted","Data":"a0dea39fcfd63a0ee62580f31d0b71f1c035a8a072b989dbf7c79ab97ba59d0c"} Dec 05 08:36:57 crc kubenswrapper[4876]: I1205 08:36:57.346108 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-876689776-4sqds" event={"ID":"579ebb91-13da-40a1-80f2-c245cb440580","Type":"ContainerStarted","Data":"d8bc27eb15aad0f3834ac0737a13d25d90d5a2dccfbbac4b9f7481199db98c6a"} Dec 05 08:36:57 crc kubenswrapper[4876]: I1205 08:36:57.348404 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-689f7d9f48-fjhvt" event={"ID":"8b5c7095-36f0-4024-b430-714d5603e08b","Type":"ContainerStarted","Data":"340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722"} Dec 05 08:36:57 crc kubenswrapper[4876]: I1205 08:36:57.348581 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:36:57 crc kubenswrapper[4876]: I1205 08:36:57.470576 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-689f7d9f48-fjhvt" podStartSLOduration=2.470551563 podStartE2EDuration="2.470551563s" podCreationTimestamp="2025-12-05 08:36:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:57.459365374 +0000 UTC m=+1281.948029996" watchObservedRunningTime="2025-12-05 08:36:57.470551563 +0000 UTC m=+1281.959216185" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.274374 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-98b56b8f5-ksg4z"] Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.276226 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.280092 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.280333 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.280384 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.307203 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-98b56b8f5-ksg4z"] Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.345106 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fbd7f2e5-b830-45be-be0b-1f23806634e2-etc-swift\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.345177 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-internal-tls-certs\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.345220 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-combined-ca-bundle\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.345270 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbd7f2e5-b830-45be-be0b-1f23806634e2-log-httpd\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.345297 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-config-data\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.345350 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbd7f2e5-b830-45be-be0b-1f23806634e2-run-httpd\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.345421 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwpqr\" (UniqueName: \"kubernetes.io/projected/fbd7f2e5-b830-45be-be0b-1f23806634e2-kube-api-access-dwpqr\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.345507 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-public-tls-certs\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.388059 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" event={"ID":"bd36d5f7-6681-4eba-a08c-28357a52cf5e","Type":"ContainerStarted","Data":"a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c"} Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.388177 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.448773 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fbd7f2e5-b830-45be-be0b-1f23806634e2-etc-swift\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.448856 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-internal-tls-certs\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.448881 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-combined-ca-bundle\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.449015 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbd7f2e5-b830-45be-be0b-1f23806634e2-log-httpd\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.449053 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-config-data\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.449098 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbd7f2e5-b830-45be-be0b-1f23806634e2-run-httpd\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.449158 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwpqr\" (UniqueName: \"kubernetes.io/projected/fbd7f2e5-b830-45be-be0b-1f23806634e2-kube-api-access-dwpqr\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.449321 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-public-tls-certs\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.452194 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbd7f2e5-b830-45be-be0b-1f23806634e2-run-httpd\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.453769 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbd7f2e5-b830-45be-be0b-1f23806634e2-log-httpd\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.459130 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fbd7f2e5-b830-45be-be0b-1f23806634e2-etc-swift\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.460807 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-config-data\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.461083 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-public-tls-certs\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.465401 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-combined-ca-bundle\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.467794 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbd7f2e5-b830-45be-be0b-1f23806634e2-internal-tls-certs\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.470694 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwpqr\" (UniqueName: \"kubernetes.io/projected/fbd7f2e5-b830-45be-be0b-1f23806634e2-kube-api-access-dwpqr\") pod \"swift-proxy-98b56b8f5-ksg4z\" (UID: \"fbd7f2e5-b830-45be-be0b-1f23806634e2\") " pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:36:58 crc kubenswrapper[4876]: I1205 08:36:58.613147 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:37:02 crc kubenswrapper[4876]: I1205 08:37:02.769579 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" podStartSLOduration=7.769554508 podStartE2EDuration="7.769554508s" podCreationTimestamp="2025-12-05 08:36:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:36:58.408107888 +0000 UTC m=+1282.896772510" watchObservedRunningTime="2025-12-05 08:37:02.769554508 +0000 UTC m=+1287.258219130" Dec 05 08:37:02 crc kubenswrapper[4876]: I1205 08:37:02.773364 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:02 crc kubenswrapper[4876]: I1205 08:37:02.773666 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="ceilometer-central-agent" containerID="cri-o://9cf53475653d7b935071f3a802715b2a47167c1a284f665854d470eb11798148" gracePeriod=30 Dec 05 08:37:02 crc kubenswrapper[4876]: I1205 08:37:02.773765 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="sg-core" containerID="cri-o://caf8ef8141c08c3c5eae3d32464d7773ae19090c1360c96e0e03ede73137bb41" gracePeriod=30 Dec 05 08:37:02 crc kubenswrapper[4876]: I1205 08:37:02.773856 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="proxy-httpd" containerID="cri-o://30eb4d1c6734bfaaf784d91bc6fe47ae09afeb912947fcf232f72666387342d2" gracePeriod=30 Dec 05 08:37:02 crc kubenswrapper[4876]: I1205 08:37:02.773868 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="ceilometer-notification-agent" containerID="cri-o://62ba626fcbf5cebe0178ed0f725dfe5c4b1921e502047ef0e1ec60a1d6e192ed" gracePeriod=30 Dec 05 08:37:02 crc kubenswrapper[4876]: I1205 08:37:02.784608 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.160:3000/\": EOF" Dec 05 08:37:03 crc kubenswrapper[4876]: I1205 08:37:03.457422 4876 generic.go:334] "Generic (PLEG): container finished" podID="f00c9009-2533-4ced-8d54-aa866361aca1" containerID="caf8ef8141c08c3c5eae3d32464d7773ae19090c1360c96e0e03ede73137bb41" exitCode=2 Dec 05 08:37:03 crc kubenswrapper[4876]: I1205 08:37:03.457469 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerDied","Data":"caf8ef8141c08c3c5eae3d32464d7773ae19090c1360c96e0e03ede73137bb41"} Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.024453 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-b64b8b465-trj6k"] Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.025689 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.046038 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-b64b8b465-trj6k"] Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.071131 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-74677b69bc-66gfk"] Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.072460 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.087893 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-74677b69bc-66gfk"] Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.110950 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-767d846448-fwbxb"] Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.112562 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.126542 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-767d846448-fwbxb"] Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.150996 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6cfq\" (UniqueName: \"kubernetes.io/projected/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-kube-api-access-j6cfq\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.152127 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-combined-ca-bundle\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.152332 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-config-data\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.152431 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-config-data-custom\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.253696 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-config-data\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.253754 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-config-data-custom\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.253813 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.253860 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-combined-ca-bundle\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.254669 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s9rd\" (UniqueName: \"kubernetes.io/projected/6c7d34c2-eacd-4cf3-840b-08ab27365769-kube-api-access-2s9rd\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.254723 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6cfq\" (UniqueName: \"kubernetes.io/projected/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-kube-api-access-j6cfq\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.254752 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6brj\" (UniqueName: \"kubernetes.io/projected/2066efee-0c4c-4e8d-913f-2cb7782b5863-kube-api-access-h6brj\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.254783 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data-custom\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.254805 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.254824 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data-custom\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.254842 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-combined-ca-bundle\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.254893 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-combined-ca-bundle\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.262919 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-combined-ca-bundle\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.266932 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-config-data-custom\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.267764 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-config-data\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.277934 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6cfq\" (UniqueName: \"kubernetes.io/projected/6a133b19-0c3f-4a85-95ea-910b9c0c4df0-kube-api-access-j6cfq\") pod \"heat-engine-b64b8b465-trj6k\" (UID: \"6a133b19-0c3f-4a85-95ea-910b9c0c4df0\") " pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.342058 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.356250 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data-custom\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.356286 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-combined-ca-bundle\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.356378 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.356402 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-combined-ca-bundle\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.356421 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s9rd\" (UniqueName: \"kubernetes.io/projected/6c7d34c2-eacd-4cf3-840b-08ab27365769-kube-api-access-2s9rd\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.356460 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6brj\" (UniqueName: \"kubernetes.io/projected/2066efee-0c4c-4e8d-913f-2cb7782b5863-kube-api-access-h6brj\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.356488 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data-custom\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.356510 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.362271 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.362857 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-combined-ca-bundle\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.363047 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data-custom\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.365517 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-combined-ca-bundle\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.365649 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data-custom\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.366883 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.373631 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s9rd\" (UniqueName: \"kubernetes.io/projected/6c7d34c2-eacd-4cf3-840b-08ab27365769-kube-api-access-2s9rd\") pod \"heat-cfnapi-767d846448-fwbxb\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.375822 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6brj\" (UniqueName: \"kubernetes.io/projected/2066efee-0c4c-4e8d-913f-2cb7782b5863-kube-api-access-h6brj\") pod \"heat-api-74677b69bc-66gfk\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.405512 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.436545 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.468779 4876 generic.go:334] "Generic (PLEG): container finished" podID="f00c9009-2533-4ced-8d54-aa866361aca1" containerID="30eb4d1c6734bfaaf784d91bc6fe47ae09afeb912947fcf232f72666387342d2" exitCode=0 Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.468807 4876 generic.go:334] "Generic (PLEG): container finished" podID="f00c9009-2533-4ced-8d54-aa866361aca1" containerID="9cf53475653d7b935071f3a802715b2a47167c1a284f665854d470eb11798148" exitCode=0 Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.468817 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerDied","Data":"30eb4d1c6734bfaaf784d91bc6fe47ae09afeb912947fcf232f72666387342d2"} Dec 05 08:37:04 crc kubenswrapper[4876]: I1205 08:37:04.468849 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerDied","Data":"9cf53475653d7b935071f3a802715b2a47167c1a284f665854d470eb11798148"} Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.498804 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5846cfbdbf-5ctlt"] Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.510232 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-876689776-4sqds"] Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.520977 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-76699577f6-nt9wb"] Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.522586 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.528099 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.530822 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.567748 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-combined-ca-bundle\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.567821 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-config-data-custom\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.568061 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-public-tls-certs\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.568161 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-config-data\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.568251 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-internal-tls-certs\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.568496 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wggn6\" (UniqueName: \"kubernetes.io/projected/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-kube-api-access-wggn6\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.571186 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5f87bc649b-spcsf"] Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.576728 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.581252 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.581706 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.599297 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-76699577f6-nt9wb"] Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.630516 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5f87bc649b-spcsf"] Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670284 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wggn6\" (UniqueName: \"kubernetes.io/projected/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-kube-api-access-wggn6\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670350 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-combined-ca-bundle\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670372 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-config-data-custom\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670394 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-public-tls-certs\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670415 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-combined-ca-bundle\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670448 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-config-data\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670477 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-public-tls-certs\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670508 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-config-data-custom\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670552 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-config-data\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670607 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-internal-tls-certs\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670629 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-internal-tls-certs\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.670659 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l4rl\" (UniqueName: \"kubernetes.io/projected/b7673120-d39b-416e-8f5f-260cdce6f71b-kube-api-access-6l4rl\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.673782 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.160:3000/\": dial tcp 10.217.0.160:3000: connect: connection refused" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.679885 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-internal-tls-certs\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.680667 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-public-tls-certs\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.684086 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-config-data\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.684231 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-config-data-custom\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.687681 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wggn6\" (UniqueName: \"kubernetes.io/projected/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-kube-api-access-wggn6\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.694705 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f3a1341-ca1e-4b8c-a57d-0c83d39a403e-combined-ca-bundle\") pod \"heat-cfnapi-76699577f6-nt9wb\" (UID: \"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e\") " pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.771836 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-public-tls-certs\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.772033 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-combined-ca-bundle\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.772143 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-config-data\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.772237 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-config-data-custom\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.772358 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-internal-tls-certs\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.772437 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l4rl\" (UniqueName: \"kubernetes.io/projected/b7673120-d39b-416e-8f5f-260cdce6f71b-kube-api-access-6l4rl\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.781232 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-combined-ca-bundle\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.784421 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.786106 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-public-tls-certs\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.786473 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-config-data\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.790722 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-config-data-custom\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.800624 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7673120-d39b-416e-8f5f-260cdce6f71b-internal-tls-certs\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.815236 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l4rl\" (UniqueName: \"kubernetes.io/projected/b7673120-d39b-416e-8f5f-260cdce6f71b-kube-api-access-6l4rl\") pod \"heat-api-5f87bc649b-spcsf\" (UID: \"b7673120-d39b-416e-8f5f-260cdce6f71b\") " pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.848126 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.866263 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2czdq"] Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.866545 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" podUID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerName="dnsmasq-dns" containerID="cri-o://1da4a1b0422093f36912f7cf0a1894e8e02880a75acd79d72d15cb5fd2faa6d9" gracePeriod=10 Dec 05 08:37:05 crc kubenswrapper[4876]: I1205 08:37:05.903467 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:06 crc kubenswrapper[4876]: I1205 08:37:06.486414 4876 generic.go:334] "Generic (PLEG): container finished" podID="f00c9009-2533-4ced-8d54-aa866361aca1" containerID="62ba626fcbf5cebe0178ed0f725dfe5c4b1921e502047ef0e1ec60a1d6e192ed" exitCode=0 Dec 05 08:37:06 crc kubenswrapper[4876]: I1205 08:37:06.486471 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerDied","Data":"62ba626fcbf5cebe0178ed0f725dfe5c4b1921e502047ef0e1ec60a1d6e192ed"} Dec 05 08:37:06 crc kubenswrapper[4876]: I1205 08:37:06.488706 4876 generic.go:334] "Generic (PLEG): container finished" podID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerID="1da4a1b0422093f36912f7cf0a1894e8e02880a75acd79d72d15cb5fd2faa6d9" exitCode=0 Dec 05 08:37:06 crc kubenswrapper[4876]: I1205 08:37:06.488752 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" event={"ID":"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f","Type":"ContainerDied","Data":"1da4a1b0422093f36912f7cf0a1894e8e02880a75acd79d72d15cb5fd2faa6d9"} Dec 05 08:37:06 crc kubenswrapper[4876]: I1205 08:37:06.860608 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" podUID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.158:5353: connect: connection refused" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.639351 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.693318 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730168 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-svc\") pod \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730255 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-sg-core-conf-yaml\") pod \"f00c9009-2533-4ced-8d54-aa866361aca1\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730313 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9tlh\" (UniqueName: \"kubernetes.io/projected/f00c9009-2533-4ced-8d54-aa866361aca1-kube-api-access-v9tlh\") pod \"f00c9009-2533-4ced-8d54-aa866361aca1\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730348 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-nb\") pod \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730386 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-scripts\") pod \"f00c9009-2533-4ced-8d54-aa866361aca1\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730412 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-combined-ca-bundle\") pod \"f00c9009-2533-4ced-8d54-aa866361aca1\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730445 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-log-httpd\") pod \"f00c9009-2533-4ced-8d54-aa866361aca1\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730477 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-sb\") pod \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730615 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmgcf\" (UniqueName: \"kubernetes.io/projected/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-kube-api-access-nmgcf\") pod \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730643 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-swift-storage-0\") pod \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730689 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-config\") pod \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\" (UID: \"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730729 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-run-httpd\") pod \"f00c9009-2533-4ced-8d54-aa866361aca1\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.730807 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-config-data\") pod \"f00c9009-2533-4ced-8d54-aa866361aca1\" (UID: \"f00c9009-2533-4ced-8d54-aa866361aca1\") " Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.740728 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f00c9009-2533-4ced-8d54-aa866361aca1" (UID: "f00c9009-2533-4ced-8d54-aa866361aca1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.741064 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-scripts" (OuterVolumeSpecName: "scripts") pod "f00c9009-2533-4ced-8d54-aa866361aca1" (UID: "f00c9009-2533-4ced-8d54-aa866361aca1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.746807 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f00c9009-2533-4ced-8d54-aa866361aca1" (UID: "f00c9009-2533-4ced-8d54-aa866361aca1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.780960 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-kube-api-access-nmgcf" (OuterVolumeSpecName: "kube-api-access-nmgcf") pod "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" (UID: "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f"). InnerVolumeSpecName "kube-api-access-nmgcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.805098 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f00c9009-2533-4ced-8d54-aa866361aca1-kube-api-access-v9tlh" (OuterVolumeSpecName: "kube-api-access-v9tlh") pod "f00c9009-2533-4ced-8d54-aa866361aca1" (UID: "f00c9009-2533-4ced-8d54-aa866361aca1"). InnerVolumeSpecName "kube-api-access-v9tlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.834734 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9tlh\" (UniqueName: \"kubernetes.io/projected/f00c9009-2533-4ced-8d54-aa866361aca1-kube-api-access-v9tlh\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.834760 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.834770 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.834781 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmgcf\" (UniqueName: \"kubernetes.io/projected/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-kube-api-access-nmgcf\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.834789 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f00c9009-2533-4ced-8d54-aa866361aca1-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.844532 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" (UID: "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.860679 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" (UID: "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.865471 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f00c9009-2533-4ced-8d54-aa866361aca1" (UID: "f00c9009-2533-4ced-8d54-aa866361aca1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.885978 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" (UID: "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.899605 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" (UID: "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.909561 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-config" (OuterVolumeSpecName: "config") pod "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" (UID: "9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.936259 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.936309 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.936325 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.936337 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.936349 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.936360 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.953009 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-config-data" (OuterVolumeSpecName: "config-data") pod "f00c9009-2533-4ced-8d54-aa866361aca1" (UID: "f00c9009-2533-4ced-8d54-aa866361aca1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:07 crc kubenswrapper[4876]: I1205 08:37:07.975213 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f00c9009-2533-4ced-8d54-aa866361aca1" (UID: "f00c9009-2533-4ced-8d54-aa866361aca1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.044234 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.044262 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f00c9009-2533-4ced-8d54-aa866361aca1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.070836 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-76699577f6-nt9wb"] Dec 05 08:37:08 crc kubenswrapper[4876]: W1205 08:37:08.101481 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f3a1341_ca1e_4b8c_a57d_0c83d39a403e.slice/crio-2acb1e64fb998c229ef157f05fc21e0385a9709720de922e0099415f661f355c WatchSource:0}: Error finding container 2acb1e64fb998c229ef157f05fc21e0385a9709720de922e0099415f661f355c: Status 404 returned error can't find the container with id 2acb1e64fb998c229ef157f05fc21e0385a9709720de922e0099415f661f355c Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.115720 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-74677b69bc-66gfk"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.214402 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.214466 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.250833 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-98b56b8f5-ksg4z"] Dec 05 08:37:08 crc kubenswrapper[4876]: W1205 08:37:08.278404 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbd7f2e5_b830_45be_be0b_1f23806634e2.slice/crio-39d06abe5d6e9dfa117e441ecceb6929277eb953ee20a8d1cf6826cb8556a9e7 WatchSource:0}: Error finding container 39d06abe5d6e9dfa117e441ecceb6929277eb953ee20a8d1cf6826cb8556a9e7: Status 404 returned error can't find the container with id 39d06abe5d6e9dfa117e441ecceb6929277eb953ee20a8d1cf6826cb8556a9e7 Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.317670 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-b64b8b465-trj6k"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.339097 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-767d846448-fwbxb"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.351736 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5f87bc649b-spcsf"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.535647 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5846cfbdbf-5ctlt" event={"ID":"718f1c24-daf5-44f1-91f3-51890c190f47","Type":"ContainerStarted","Data":"50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.535874 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5846cfbdbf-5ctlt" podUID="718f1c24-daf5-44f1-91f3-51890c190f47" containerName="heat-api" containerID="cri-o://50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6" gracePeriod=60 Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.536622 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.548464 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"86ac21be-c7c2-4035-a32c-3f24715c3248","Type":"ContainerStarted","Data":"48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.567999 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f00c9009-2533-4ced-8d54-aa866361aca1","Type":"ContainerDied","Data":"726d7c8d8955e7c20391f73c73d55748ed3dffdcd12dcb24b52d1df4d6c0ff72"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.568166 4876 scope.go:117] "RemoveContainer" containerID="30eb4d1c6734bfaaf784d91bc6fe47ae09afeb912947fcf232f72666387342d2" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.568032 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.575193 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-74677b69bc-66gfk" event={"ID":"2066efee-0c4c-4e8d-913f-2cb7782b5863","Type":"ContainerStarted","Data":"7fc0c5bb28b3208805bf722b6aae97b70704b8c30bb18ebcce3d8d30cddddbfa"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.575244 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-74677b69bc-66gfk" event={"ID":"2066efee-0c4c-4e8d-913f-2cb7782b5863","Type":"ContainerStarted","Data":"af0445a57870ffa94bf5a3251a1e6cab22c8a01ca5d944b3d04c610b4433b596"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.575868 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.579956 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.324395469 podStartE2EDuration="16.579941038s" podCreationTimestamp="2025-12-05 08:36:52 +0000 UTC" firstStartedPulling="2025-12-05 08:36:53.000821615 +0000 UTC m=+1277.489486257" lastFinishedPulling="2025-12-05 08:37:07.256367194 +0000 UTC m=+1291.745031826" observedRunningTime="2025-12-05 08:37:08.579143785 +0000 UTC m=+1293.067808407" watchObservedRunningTime="2025-12-05 08:37:08.579941038 +0000 UTC m=+1293.068605650" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.583590 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.587076 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5846cfbdbf-5ctlt" podStartSLOduration=3.328118704 podStartE2EDuration="13.587057341s" podCreationTimestamp="2025-12-05 08:36:55 +0000 UTC" firstStartedPulling="2025-12-05 08:36:56.780767427 +0000 UTC m=+1281.269432049" lastFinishedPulling="2025-12-05 08:37:07.039706064 +0000 UTC m=+1291.528370686" observedRunningTime="2025-12-05 08:37:08.56245925 +0000 UTC m=+1293.051123872" watchObservedRunningTime="2025-12-05 08:37:08.587057341 +0000 UTC m=+1293.075721973" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.583496 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-2czdq" event={"ID":"9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f","Type":"ContainerDied","Data":"b00c4086cef3bbdd3bba6563f5785420ce514cdd89a0398e2d83f02f107e62ce"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.595964 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-98b56b8f5-ksg4z" event={"ID":"fbd7f2e5-b830-45be-be0b-1f23806634e2","Type":"ContainerStarted","Data":"39d06abe5d6e9dfa117e441ecceb6929277eb953ee20a8d1cf6826cb8556a9e7"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.598469 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5f87bc649b-spcsf" event={"ID":"b7673120-d39b-416e-8f5f-260cdce6f71b","Type":"ContainerStarted","Data":"7112a506792d034541318317714ea12f67f27745333fc5c629481f4dc1cf790e"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.605664 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-876689776-4sqds" event={"ID":"579ebb91-13da-40a1-80f2-c245cb440580","Type":"ContainerStarted","Data":"4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.605757 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-876689776-4sqds" podUID="579ebb91-13da-40a1-80f2-c245cb440580" containerName="heat-cfnapi" containerID="cri-o://4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749" gracePeriod=60 Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.605792 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.607334 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-74677b69bc-66gfk" podStartSLOduration=4.607323629 podStartE2EDuration="4.607323629s" podCreationTimestamp="2025-12-05 08:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:08.598134117 +0000 UTC m=+1293.086798739" watchObservedRunningTime="2025-12-05 08:37:08.607323629 +0000 UTC m=+1293.095988251" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.611005 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-b64b8b465-trj6k" event={"ID":"6a133b19-0c3f-4a85-95ea-910b9c0c4df0","Type":"ContainerStarted","Data":"358e228032ed82190aa9124119870965cf37c66a6d71cda47255409b6eb9781e"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.619648 4876 scope.go:117] "RemoveContainer" containerID="caf8ef8141c08c3c5eae3d32464d7773ae19090c1360c96e0e03ede73137bb41" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.619772 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-767d846448-fwbxb" event={"ID":"6c7d34c2-eacd-4cf3-840b-08ab27365769","Type":"ContainerStarted","Data":"2444ae922e59f4a83523fec1f189d2b24386c4571d2ba2f3507046877ee28c40"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.635671 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-76699577f6-nt9wb" event={"ID":"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e","Type":"ContainerStarted","Data":"d866f0255d557d4a4f0cc94332baaca80316dd8c7015181c30ef3618084e689f"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.635711 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-76699577f6-nt9wb" event={"ID":"0f3a1341-ca1e-4b8c-a57d-0c83d39a403e","Type":"ContainerStarted","Data":"2acb1e64fb998c229ef157f05fc21e0385a9709720de922e0099415f661f355c"} Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.636817 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.671679 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-876689776-4sqds" podStartSLOduration=3.434256743 podStartE2EDuration="13.671663195s" podCreationTimestamp="2025-12-05 08:36:55 +0000 UTC" firstStartedPulling="2025-12-05 08:36:56.716072732 +0000 UTC m=+1281.204737354" lastFinishedPulling="2025-12-05 08:37:06.953479194 +0000 UTC m=+1291.442143806" observedRunningTime="2025-12-05 08:37:08.631731545 +0000 UTC m=+1293.120396167" watchObservedRunningTime="2025-12-05 08:37:08.671663195 +0000 UTC m=+1293.160327817" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.687474 4876 scope.go:117] "RemoveContainer" containerID="62ba626fcbf5cebe0178ed0f725dfe5c4b1921e502047ef0e1ec60a1d6e192ed" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.698444 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.723806 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-76699577f6-nt9wb" podStartSLOduration=3.7237902309999997 podStartE2EDuration="3.723790231s" podCreationTimestamp="2025-12-05 08:37:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:08.6806121 +0000 UTC m=+1293.169276722" watchObservedRunningTime="2025-12-05 08:37:08.723790231 +0000 UTC m=+1293.212454853" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.733104 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755034 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:08 crc kubenswrapper[4876]: E1205 08:37:08.755549 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="sg-core" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755575 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="sg-core" Dec 05 08:37:08 crc kubenswrapper[4876]: E1205 08:37:08.755592 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerName="dnsmasq-dns" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755600 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerName="dnsmasq-dns" Dec 05 08:37:08 crc kubenswrapper[4876]: E1205 08:37:08.755613 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerName="init" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755621 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerName="init" Dec 05 08:37:08 crc kubenswrapper[4876]: E1205 08:37:08.755639 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="ceilometer-notification-agent" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755647 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="ceilometer-notification-agent" Dec 05 08:37:08 crc kubenswrapper[4876]: E1205 08:37:08.755669 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="ceilometer-central-agent" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755679 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="ceilometer-central-agent" Dec 05 08:37:08 crc kubenswrapper[4876]: E1205 08:37:08.755703 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="proxy-httpd" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755711 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="proxy-httpd" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755963 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" containerName="dnsmasq-dns" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755978 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="ceilometer-central-agent" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.755995 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="proxy-httpd" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.756013 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="ceilometer-notification-agent" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.756033 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" containerName="sg-core" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.758317 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.760955 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2czdq"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.769136 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.769331 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.772183 4876 scope.go:117] "RemoveContainer" containerID="9cf53475653d7b935071f3a802715b2a47167c1a284f665854d470eb11798148" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.772445 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-2czdq"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.791672 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.823127 4876 scope.go:117] "RemoveContainer" containerID="1da4a1b0422093f36912f7cf0a1894e8e02880a75acd79d72d15cb5fd2faa6d9" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.864733 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-config-data\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.864846 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-log-httpd\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.864869 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9zvv\" (UniqueName: \"kubernetes.io/projected/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-kube-api-access-f9zvv\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.864890 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-run-httpd\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.864952 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.864974 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-scripts\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.865000 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.966127 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-config-data\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.966247 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-log-httpd\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.966270 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9zvv\" (UniqueName: \"kubernetes.io/projected/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-kube-api-access-f9zvv\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.966294 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-run-httpd\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.966328 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.966352 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-scripts\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.966380 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.968503 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-run-httpd\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.968716 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-log-httpd\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.973978 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.976605 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.977823 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-config-data\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.982790 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-scripts\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:08 crc kubenswrapper[4876]: I1205 08:37:08.983577 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9zvv\" (UniqueName: \"kubernetes.io/projected/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-kube-api-access-f9zvv\") pod \"ceilometer-0\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " pod="openstack/ceilometer-0" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.082330 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.257772 4876 scope.go:117] "RemoveContainer" containerID="ee8c99d68773c53b0b065da2715f54049a76de9a415e0b9d3eca801dd14691ca" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.582277 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.597459 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.615685 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cvjh\" (UniqueName: \"kubernetes.io/projected/579ebb91-13da-40a1-80f2-c245cb440580-kube-api-access-5cvjh\") pod \"579ebb91-13da-40a1-80f2-c245cb440580\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.615953 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-combined-ca-bundle\") pod \"579ebb91-13da-40a1-80f2-c245cb440580\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.616109 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data\") pod \"579ebb91-13da-40a1-80f2-c245cb440580\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.616203 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data\") pod \"718f1c24-daf5-44f1-91f3-51890c190f47\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.616306 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data-custom\") pod \"579ebb91-13da-40a1-80f2-c245cb440580\" (UID: \"579ebb91-13da-40a1-80f2-c245cb440580\") " Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.616394 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data-custom\") pod \"718f1c24-daf5-44f1-91f3-51890c190f47\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.616488 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvpd2\" (UniqueName: \"kubernetes.io/projected/718f1c24-daf5-44f1-91f3-51890c190f47-kube-api-access-vvpd2\") pod \"718f1c24-daf5-44f1-91f3-51890c190f47\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.616573 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-combined-ca-bundle\") pod \"718f1c24-daf5-44f1-91f3-51890c190f47\" (UID: \"718f1c24-daf5-44f1-91f3-51890c190f47\") " Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.617872 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.632202 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "718f1c24-daf5-44f1-91f3-51890c190f47" (UID: "718f1c24-daf5-44f1-91f3-51890c190f47"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.634202 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "579ebb91-13da-40a1-80f2-c245cb440580" (UID: "579ebb91-13da-40a1-80f2-c245cb440580"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.636347 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/579ebb91-13da-40a1-80f2-c245cb440580-kube-api-access-5cvjh" (OuterVolumeSpecName: "kube-api-access-5cvjh") pod "579ebb91-13da-40a1-80f2-c245cb440580" (UID: "579ebb91-13da-40a1-80f2-c245cb440580"). InnerVolumeSpecName "kube-api-access-5cvjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.641723 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/718f1c24-daf5-44f1-91f3-51890c190f47-kube-api-access-vvpd2" (OuterVolumeSpecName: "kube-api-access-vvpd2") pod "718f1c24-daf5-44f1-91f3-51890c190f47" (UID: "718f1c24-daf5-44f1-91f3-51890c190f47"). InnerVolumeSpecName "kube-api-access-vvpd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.647661 4876 generic.go:334] "Generic (PLEG): container finished" podID="2066efee-0c4c-4e8d-913f-2cb7782b5863" containerID="7fc0c5bb28b3208805bf722b6aae97b70704b8c30bb18ebcce3d8d30cddddbfa" exitCode=1 Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.647718 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-74677b69bc-66gfk" event={"ID":"2066efee-0c4c-4e8d-913f-2cb7782b5863","Type":"ContainerDied","Data":"7fc0c5bb28b3208805bf722b6aae97b70704b8c30bb18ebcce3d8d30cddddbfa"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.648395 4876 scope.go:117] "RemoveContainer" containerID="7fc0c5bb28b3208805bf722b6aae97b70704b8c30bb18ebcce3d8d30cddddbfa" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.671302 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "579ebb91-13da-40a1-80f2-c245cb440580" (UID: "579ebb91-13da-40a1-80f2-c245cb440580"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.691080 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data" (OuterVolumeSpecName: "config-data") pod "718f1c24-daf5-44f1-91f3-51890c190f47" (UID: "718f1c24-daf5-44f1-91f3-51890c190f47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.694943 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5f87bc649b-spcsf" event={"ID":"b7673120-d39b-416e-8f5f-260cdce6f71b","Type":"ContainerStarted","Data":"c6b08919a4f68bf02b593397d5ffbc658a8a4e98af41d0736b250dae2b5a87b7"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.695048 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.699814 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "718f1c24-daf5-44f1-91f3-51890c190f47" (UID: "718f1c24-daf5-44f1-91f3-51890c190f47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.705684 4876 generic.go:334] "Generic (PLEG): container finished" podID="718f1c24-daf5-44f1-91f3-51890c190f47" containerID="50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6" exitCode=0 Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.705796 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5846cfbdbf-5ctlt" event={"ID":"718f1c24-daf5-44f1-91f3-51890c190f47","Type":"ContainerDied","Data":"50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.705832 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5846cfbdbf-5ctlt" event={"ID":"718f1c24-daf5-44f1-91f3-51890c190f47","Type":"ContainerDied","Data":"a0dea39fcfd63a0ee62580f31d0b71f1c035a8a072b989dbf7c79ab97ba59d0c"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.705854 4876 scope.go:117] "RemoveContainer" containerID="50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.706135 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5846cfbdbf-5ctlt" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.715587 4876 generic.go:334] "Generic (PLEG): container finished" podID="579ebb91-13da-40a1-80f2-c245cb440580" containerID="4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749" exitCode=0 Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.715681 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-876689776-4sqds" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.716078 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-876689776-4sqds" event={"ID":"579ebb91-13da-40a1-80f2-c245cb440580","Type":"ContainerDied","Data":"4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.716108 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-876689776-4sqds" event={"ID":"579ebb91-13da-40a1-80f2-c245cb440580","Type":"ContainerDied","Data":"d8bc27eb15aad0f3834ac0737a13d25d90d5a2dccfbbac4b9f7481199db98c6a"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.725179 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.725202 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.725211 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvpd2\" (UniqueName: \"kubernetes.io/projected/718f1c24-daf5-44f1-91f3-51890c190f47-kube-api-access-vvpd2\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.725221 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.725231 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cvjh\" (UniqueName: \"kubernetes.io/projected/579ebb91-13da-40a1-80f2-c245cb440580-kube-api-access-5cvjh\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.725240 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.725250 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/718f1c24-daf5-44f1-91f3-51890c190f47-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.727071 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5f87bc649b-spcsf" podStartSLOduration=4.727060308 podStartE2EDuration="4.727060308s" podCreationTimestamp="2025-12-05 08:37:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:09.716972432 +0000 UTC m=+1294.205637054" watchObservedRunningTime="2025-12-05 08:37:09.727060308 +0000 UTC m=+1294.215724930" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.729917 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-b64b8b465-trj6k" event={"ID":"6a133b19-0c3f-4a85-95ea-910b9c0c4df0","Type":"ContainerStarted","Data":"e50a42cc2349f81bf62cd45813f43c4431cbd3d8a9f93ecd0de17ef5781bee37"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.730778 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.740267 4876 generic.go:334] "Generic (PLEG): container finished" podID="6c7d34c2-eacd-4cf3-840b-08ab27365769" containerID="0c6cfcd3e103a757fb67d3d34c259e6bfb256b4217f10d3cd618613563308ea0" exitCode=1 Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.741161 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-767d846448-fwbxb" event={"ID":"6c7d34c2-eacd-4cf3-840b-08ab27365769","Type":"ContainerDied","Data":"0c6cfcd3e103a757fb67d3d34c259e6bfb256b4217f10d3cd618613563308ea0"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.740890 4876 scope.go:117] "RemoveContainer" containerID="0c6cfcd3e103a757fb67d3d34c259e6bfb256b4217f10d3cd618613563308ea0" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.744858 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-98b56b8f5-ksg4z" event={"ID":"fbd7f2e5-b830-45be-be0b-1f23806634e2","Type":"ContainerStarted","Data":"5a9b12c217e4806078c13ee73630b120ff4c61cf73686a42312a0f19ceacc864"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.744921 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-98b56b8f5-ksg4z" event={"ID":"fbd7f2e5-b830-45be-be0b-1f23806634e2","Type":"ContainerStarted","Data":"9d901942402fd2069bbdced7998403490ec0887f56a101e125f39689bd95694d"} Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.745062 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.745076 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.753079 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-b64b8b465-trj6k" podStartSLOduration=6.753059673 podStartE2EDuration="6.753059673s" podCreationTimestamp="2025-12-05 08:37:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:09.751634425 +0000 UTC m=+1294.240299047" watchObservedRunningTime="2025-12-05 08:37:09.753059673 +0000 UTC m=+1294.241724295" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.754019 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data" (OuterVolumeSpecName: "config-data") pod "579ebb91-13da-40a1-80f2-c245cb440580" (UID: "579ebb91-13da-40a1-80f2-c245cb440580"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.761015 4876 scope.go:117] "RemoveContainer" containerID="50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6" Dec 05 08:37:09 crc kubenswrapper[4876]: E1205 08:37:09.763344 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6\": container with ID starting with 50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6 not found: ID does not exist" containerID="50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.763382 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6"} err="failed to get container status \"50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6\": rpc error: code = NotFound desc = could not find container \"50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6\": container with ID starting with 50e33fcd059d5ba8e800aa89ab2a66fbd21cf806c1ebd2c1f5dc0bc8253721c6 not found: ID does not exist" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.763404 4876 scope.go:117] "RemoveContainer" containerID="4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.789524 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5846cfbdbf-5ctlt"] Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.805858 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5846cfbdbf-5ctlt"] Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.828827 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/579ebb91-13da-40a1-80f2-c245cb440580-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.833163 4876 scope.go:117] "RemoveContainer" containerID="4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749" Dec 05 08:37:09 crc kubenswrapper[4876]: E1205 08:37:09.833696 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749\": container with ID starting with 4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749 not found: ID does not exist" containerID="4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.833731 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749"} err="failed to get container status \"4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749\": rpc error: code = NotFound desc = could not find container \"4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749\": container with ID starting with 4e30e49936dc81eb0c8520cd2ced9dd1615d1f3a7f13d0f2296d2a1c81313749 not found: ID does not exist" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.837070 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="718f1c24-daf5-44f1-91f3-51890c190f47" path="/var/lib/kubelet/pods/718f1c24-daf5-44f1-91f3-51890c190f47/volumes" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.837817 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f" path="/var/lib/kubelet/pods/9aca26c9-2dd2-46bb-ad3a-4b5e5bb4e07f/volumes" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.838839 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f00c9009-2533-4ced-8d54-aa866361aca1" path="/var/lib/kubelet/pods/f00c9009-2533-4ced-8d54-aa866361aca1/volumes" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.847696 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-98b56b8f5-ksg4z" podStartSLOduration=11.847678975000001 podStartE2EDuration="11.847678975s" podCreationTimestamp="2025-12-05 08:36:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:09.81257265 +0000 UTC m=+1294.301237262" watchObservedRunningTime="2025-12-05 08:37:09.847678975 +0000 UTC m=+1294.336343597" Dec 05 08:37:09 crc kubenswrapper[4876]: I1205 08:37:09.884200 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.061607 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-876689776-4sqds"] Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.068202 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-876689776-4sqds"] Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.782266 4876 generic.go:334] "Generic (PLEG): container finished" podID="6c7d34c2-eacd-4cf3-840b-08ab27365769" containerID="d92f1b42a0ff4c27934987cf8c1c33b97a985461e67bc370a23a26cc43f61e32" exitCode=1 Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.782321 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-767d846448-fwbxb" event={"ID":"6c7d34c2-eacd-4cf3-840b-08ab27365769","Type":"ContainerDied","Data":"d92f1b42a0ff4c27934987cf8c1c33b97a985461e67bc370a23a26cc43f61e32"} Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.782601 4876 scope.go:117] "RemoveContainer" containerID="0c6cfcd3e103a757fb67d3d34c259e6bfb256b4217f10d3cd618613563308ea0" Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.783228 4876 scope.go:117] "RemoveContainer" containerID="d92f1b42a0ff4c27934987cf8c1c33b97a985461e67bc370a23a26cc43f61e32" Dec 05 08:37:10 crc kubenswrapper[4876]: E1205 08:37:10.783617 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-767d846448-fwbxb_openstack(6c7d34c2-eacd-4cf3-840b-08ab27365769)\"" pod="openstack/heat-cfnapi-767d846448-fwbxb" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.833456 4876 generic.go:334] "Generic (PLEG): container finished" podID="2066efee-0c4c-4e8d-913f-2cb7782b5863" containerID="cb3715ab52291aa14e991b69ebb5e5514f25842602da4b54043af15ed7dd7401" exitCode=1 Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.833804 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-74677b69bc-66gfk" event={"ID":"2066efee-0c4c-4e8d-913f-2cb7782b5863","Type":"ContainerDied","Data":"cb3715ab52291aa14e991b69ebb5e5514f25842602da4b54043af15ed7dd7401"} Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.834201 4876 scope.go:117] "RemoveContainer" containerID="cb3715ab52291aa14e991b69ebb5e5514f25842602da4b54043af15ed7dd7401" Dec 05 08:37:10 crc kubenswrapper[4876]: E1205 08:37:10.834463 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-74677b69bc-66gfk_openstack(2066efee-0c4c-4e8d-913f-2cb7782b5863)\"" pod="openstack/heat-api-74677b69bc-66gfk" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.849968 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerStarted","Data":"282186d79c12f24f859b943d17c1a42f3ec95ee5a69a9a7c212c1b6b129d6d66"} Dec 05 08:37:10 crc kubenswrapper[4876]: I1205 08:37:10.850037 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerStarted","Data":"c9b24e08a18a0afd5a75242486bcc0a8670b99dcb5b9af75d5684350c785013f"} Dec 05 08:37:11 crc kubenswrapper[4876]: I1205 08:37:11.008189 4876 scope.go:117] "RemoveContainer" containerID="7fc0c5bb28b3208805bf722b6aae97b70704b8c30bb18ebcce3d8d30cddddbfa" Dec 05 08:37:11 crc kubenswrapper[4876]: I1205 08:37:11.836063 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="579ebb91-13da-40a1-80f2-c245cb440580" path="/var/lib/kubelet/pods/579ebb91-13da-40a1-80f2-c245cb440580/volumes" Dec 05 08:37:11 crc kubenswrapper[4876]: I1205 08:37:11.860458 4876 scope.go:117] "RemoveContainer" containerID="cb3715ab52291aa14e991b69ebb5e5514f25842602da4b54043af15ed7dd7401" Dec 05 08:37:11 crc kubenswrapper[4876]: E1205 08:37:11.860684 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-74677b69bc-66gfk_openstack(2066efee-0c4c-4e8d-913f-2cb7782b5863)\"" pod="openstack/heat-api-74677b69bc-66gfk" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" Dec 05 08:37:11 crc kubenswrapper[4876]: I1205 08:37:11.863301 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerStarted","Data":"6d2cd6ea6447c668eb5814192932fc05d8c0025eae5bb579833725ff82bbecc7"} Dec 05 08:37:11 crc kubenswrapper[4876]: I1205 08:37:11.865790 4876 scope.go:117] "RemoveContainer" containerID="d92f1b42a0ff4c27934987cf8c1c33b97a985461e67bc370a23a26cc43f61e32" Dec 05 08:37:11 crc kubenswrapper[4876]: E1205 08:37:11.866028 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-767d846448-fwbxb_openstack(6c7d34c2-eacd-4cf3-840b-08ab27365769)\"" pod="openstack/heat-cfnapi-767d846448-fwbxb" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.182355 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hb5wk"] Dec 05 08:37:12 crc kubenswrapper[4876]: E1205 08:37:12.183032 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="718f1c24-daf5-44f1-91f3-51890c190f47" containerName="heat-api" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.183216 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="718f1c24-daf5-44f1-91f3-51890c190f47" containerName="heat-api" Dec 05 08:37:12 crc kubenswrapper[4876]: E1205 08:37:12.183311 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="579ebb91-13da-40a1-80f2-c245cb440580" containerName="heat-cfnapi" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.183412 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="579ebb91-13da-40a1-80f2-c245cb440580" containerName="heat-cfnapi" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.183784 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="579ebb91-13da-40a1-80f2-c245cb440580" containerName="heat-cfnapi" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.183868 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="718f1c24-daf5-44f1-91f3-51890c190f47" containerName="heat-api" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.184580 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.198721 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hb5wk"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.282142 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-z22v8"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.283739 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.293199 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70aa4ff-ac52-4cef-808e-6133ef7d395f-operator-scripts\") pod \"nova-api-db-create-hb5wk\" (UID: \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\") " pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.293326 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nccfg\" (UniqueName: \"kubernetes.io/projected/f70aa4ff-ac52-4cef-808e-6133ef7d395f-kube-api-access-nccfg\") pod \"nova-api-db-create-hb5wk\" (UID: \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\") " pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.297345 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z22v8"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.309427 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-d661-account-create-update-cb7c4"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.310773 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.330245 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.343322 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d661-account-create-update-cb7c4"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.395220 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh9x2\" (UniqueName: \"kubernetes.io/projected/ed3eafa5-4131-4235-aba0-de93198af37d-kube-api-access-xh9x2\") pod \"nova-api-d661-account-create-update-cb7c4\" (UID: \"ed3eafa5-4131-4235-aba0-de93198af37d\") " pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.395304 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76pdw\" (UniqueName: \"kubernetes.io/projected/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-kube-api-access-76pdw\") pod \"nova-cell0-db-create-z22v8\" (UID: \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\") " pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.395358 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-operator-scripts\") pod \"nova-cell0-db-create-z22v8\" (UID: \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\") " pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.395458 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3eafa5-4131-4235-aba0-de93198af37d-operator-scripts\") pod \"nova-api-d661-account-create-update-cb7c4\" (UID: \"ed3eafa5-4131-4235-aba0-de93198af37d\") " pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.395517 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70aa4ff-ac52-4cef-808e-6133ef7d395f-operator-scripts\") pod \"nova-api-db-create-hb5wk\" (UID: \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\") " pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.395624 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nccfg\" (UniqueName: \"kubernetes.io/projected/f70aa4ff-ac52-4cef-808e-6133ef7d395f-kube-api-access-nccfg\") pod \"nova-api-db-create-hb5wk\" (UID: \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\") " pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.396752 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70aa4ff-ac52-4cef-808e-6133ef7d395f-operator-scripts\") pod \"nova-api-db-create-hb5wk\" (UID: \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\") " pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.428963 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-9cljp"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.430495 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.441694 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9cljp"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.460583 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nccfg\" (UniqueName: \"kubernetes.io/projected/f70aa4ff-ac52-4cef-808e-6133ef7d395f-kube-api-access-nccfg\") pod \"nova-api-db-create-hb5wk\" (UID: \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\") " pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.498801 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh9x2\" (UniqueName: \"kubernetes.io/projected/ed3eafa5-4131-4235-aba0-de93198af37d-kube-api-access-xh9x2\") pod \"nova-api-d661-account-create-update-cb7c4\" (UID: \"ed3eafa5-4131-4235-aba0-de93198af37d\") " pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.498884 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76pdw\" (UniqueName: \"kubernetes.io/projected/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-kube-api-access-76pdw\") pod \"nova-cell0-db-create-z22v8\" (UID: \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\") " pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.498948 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sw8j\" (UniqueName: \"kubernetes.io/projected/1e638147-a228-4f35-b3f1-8ead1cc5db48-kube-api-access-7sw8j\") pod \"nova-cell1-db-create-9cljp\" (UID: \"1e638147-a228-4f35-b3f1-8ead1cc5db48\") " pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.498988 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-operator-scripts\") pod \"nova-cell0-db-create-z22v8\" (UID: \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\") " pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.499062 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3eafa5-4131-4235-aba0-de93198af37d-operator-scripts\") pod \"nova-api-d661-account-create-update-cb7c4\" (UID: \"ed3eafa5-4131-4235-aba0-de93198af37d\") " pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.499104 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e638147-a228-4f35-b3f1-8ead1cc5db48-operator-scripts\") pod \"nova-cell1-db-create-9cljp\" (UID: \"1e638147-a228-4f35-b3f1-8ead1cc5db48\") " pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.500715 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-operator-scripts\") pod \"nova-cell0-db-create-z22v8\" (UID: \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\") " pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.501385 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3eafa5-4131-4235-aba0-de93198af37d-operator-scripts\") pod \"nova-api-d661-account-create-update-cb7c4\" (UID: \"ed3eafa5-4131-4235-aba0-de93198af37d\") " pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.507834 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.513961 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-87f7-account-create-update-fsgxj"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.515502 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.519226 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.530527 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh9x2\" (UniqueName: \"kubernetes.io/projected/ed3eafa5-4131-4235-aba0-de93198af37d-kube-api-access-xh9x2\") pod \"nova-api-d661-account-create-update-cb7c4\" (UID: \"ed3eafa5-4131-4235-aba0-de93198af37d\") " pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.555971 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-87f7-account-create-update-fsgxj"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.579949 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76pdw\" (UniqueName: \"kubernetes.io/projected/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-kube-api-access-76pdw\") pod \"nova-cell0-db-create-z22v8\" (UID: \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\") " pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.600591 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e638147-a228-4f35-b3f1-8ead1cc5db48-operator-scripts\") pod \"nova-cell1-db-create-9cljp\" (UID: \"1e638147-a228-4f35-b3f1-8ead1cc5db48\") " pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.609652 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e638147-a228-4f35-b3f1-8ead1cc5db48-operator-scripts\") pod \"nova-cell1-db-create-9cljp\" (UID: \"1e638147-a228-4f35-b3f1-8ead1cc5db48\") " pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.609810 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sw8j\" (UniqueName: \"kubernetes.io/projected/1e638147-a228-4f35-b3f1-8ead1cc5db48-kube-api-access-7sw8j\") pod \"nova-cell1-db-create-9cljp\" (UID: \"1e638147-a228-4f35-b3f1-8ead1cc5db48\") " pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.610314 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.637723 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.644541 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sw8j\" (UniqueName: \"kubernetes.io/projected/1e638147-a228-4f35-b3f1-8ead1cc5db48-kube-api-access-7sw8j\") pod \"nova-cell1-db-create-9cljp\" (UID: \"1e638147-a228-4f35-b3f1-8ead1cc5db48\") " pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.696721 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-b05e-account-create-update-5t4rh"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.703297 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.711092 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.711796 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bce654-ab4b-4687-8216-21761f83a696-operator-scripts\") pod \"nova-cell0-87f7-account-create-update-fsgxj\" (UID: \"d2bce654-ab4b-4687-8216-21761f83a696\") " pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.711968 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fp98\" (UniqueName: \"kubernetes.io/projected/d2bce654-ab4b-4687-8216-21761f83a696-kube-api-access-6fp98\") pod \"nova-cell0-87f7-account-create-update-fsgxj\" (UID: \"d2bce654-ab4b-4687-8216-21761f83a696\") " pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.712693 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b05e-account-create-update-5t4rh"] Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.797755 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.849374 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/693e7fd3-fc4e-44fe-827c-2063e9d2f496-operator-scripts\") pod \"nova-cell1-b05e-account-create-update-5t4rh\" (UID: \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\") " pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.849499 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fp98\" (UniqueName: \"kubernetes.io/projected/d2bce654-ab4b-4687-8216-21761f83a696-kube-api-access-6fp98\") pod \"nova-cell0-87f7-account-create-update-fsgxj\" (UID: \"d2bce654-ab4b-4687-8216-21761f83a696\") " pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.849601 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcfnt\" (UniqueName: \"kubernetes.io/projected/693e7fd3-fc4e-44fe-827c-2063e9d2f496-kube-api-access-fcfnt\") pod \"nova-cell1-b05e-account-create-update-5t4rh\" (UID: \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\") " pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.849667 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bce654-ab4b-4687-8216-21761f83a696-operator-scripts\") pod \"nova-cell0-87f7-account-create-update-fsgxj\" (UID: \"d2bce654-ab4b-4687-8216-21761f83a696\") " pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.851301 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bce654-ab4b-4687-8216-21761f83a696-operator-scripts\") pod \"nova-cell0-87f7-account-create-update-fsgxj\" (UID: \"d2bce654-ab4b-4687-8216-21761f83a696\") " pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.879050 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fp98\" (UniqueName: \"kubernetes.io/projected/d2bce654-ab4b-4687-8216-21761f83a696-kube-api-access-6fp98\") pod \"nova-cell0-87f7-account-create-update-fsgxj\" (UID: \"d2bce654-ab4b-4687-8216-21761f83a696\") " pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.887100 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerStarted","Data":"a486f16070c5ab7dc259453d73a209740231be7646a624574b751659ed9b079c"} Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.953336 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/693e7fd3-fc4e-44fe-827c-2063e9d2f496-operator-scripts\") pod \"nova-cell1-b05e-account-create-update-5t4rh\" (UID: \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\") " pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.954044 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcfnt\" (UniqueName: \"kubernetes.io/projected/693e7fd3-fc4e-44fe-827c-2063e9d2f496-kube-api-access-fcfnt\") pod \"nova-cell1-b05e-account-create-update-5t4rh\" (UID: \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\") " pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.956869 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/693e7fd3-fc4e-44fe-827c-2063e9d2f496-operator-scripts\") pod \"nova-cell1-b05e-account-create-update-5t4rh\" (UID: \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\") " pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.959653 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:12 crc kubenswrapper[4876]: I1205 08:37:12.974655 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcfnt\" (UniqueName: \"kubernetes.io/projected/693e7fd3-fc4e-44fe-827c-2063e9d2f496-kube-api-access-fcfnt\") pod \"nova-cell1-b05e-account-create-update-5t4rh\" (UID: \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\") " pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.103419 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.245919 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z22v8"] Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.313714 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hb5wk"] Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.432615 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-d661-account-create-update-cb7c4"] Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.620105 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-87f7-account-create-update-fsgxj"] Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.653978 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9cljp"] Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.658890 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:37:13 crc kubenswrapper[4876]: W1205 08:37:13.820420 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod693e7fd3_fc4e_44fe_827c_2063e9d2f496.slice/crio-19c4e5715430a633b8c0be21e66e431c4ef682437d84041c63e2a6084a4106a4 WatchSource:0}: Error finding container 19c4e5715430a633b8c0be21e66e431c4ef682437d84041c63e2a6084a4106a4: Status 404 returned error can't find the container with id 19c4e5715430a633b8c0be21e66e431c4ef682437d84041c63e2a6084a4106a4 Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.834566 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b05e-account-create-update-5t4rh"] Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.898711 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hb5wk" event={"ID":"f70aa4ff-ac52-4cef-808e-6133ef7d395f","Type":"ContainerStarted","Data":"ee6a584c12138550fdb9b7e08cc4c6fe4521234f899036b256c27465d360c351"} Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.900174 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" event={"ID":"693e7fd3-fc4e-44fe-827c-2063e9d2f496","Type":"ContainerStarted","Data":"19c4e5715430a633b8c0be21e66e431c4ef682437d84041c63e2a6084a4106a4"} Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.901193 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z22v8" event={"ID":"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3","Type":"ContainerStarted","Data":"52b3a4f3b2c53c6e2a6471a3ddb38355dd09647c35ad0df039ab173b7693271b"} Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.902071 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9cljp" event={"ID":"1e638147-a228-4f35-b3f1-8ead1cc5db48","Type":"ContainerStarted","Data":"071624cd9f8b7f6e72d55c4da060c4142f3567e5c53ef1a2adf768718fc312b0"} Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.904497 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d661-account-create-update-cb7c4" event={"ID":"ed3eafa5-4131-4235-aba0-de93198af37d","Type":"ContainerStarted","Data":"9d97da2f6d71e512b99ff23d5d6b739f205d5c1ca6a6106ea2944ac4b6bbcd5f"} Dec 05 08:37:13 crc kubenswrapper[4876]: I1205 08:37:13.905791 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" event={"ID":"d2bce654-ab4b-4687-8216-21761f83a696","Type":"ContainerStarted","Data":"ce4482fff8dcf003080d40a22b232ef255b6110489ab9a16c1283aabe4317117"} Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.406522 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.406839 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.407607 4876 scope.go:117] "RemoveContainer" containerID="cb3715ab52291aa14e991b69ebb5e5514f25842602da4b54043af15ed7dd7401" Dec 05 08:37:14 crc kubenswrapper[4876]: E1205 08:37:14.407922 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-74677b69bc-66gfk_openstack(2066efee-0c4c-4e8d-913f-2cb7782b5863)\"" pod="openstack/heat-api-74677b69bc-66gfk" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.437332 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.437392 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.438168 4876 scope.go:117] "RemoveContainer" containerID="d92f1b42a0ff4c27934987cf8c1c33b97a985461e67bc370a23a26cc43f61e32" Dec 05 08:37:14 crc kubenswrapper[4876]: E1205 08:37:14.438424 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-767d846448-fwbxb_openstack(6c7d34c2-eacd-4cf3-840b-08ab27365769)\"" pod="openstack/heat-cfnapi-767d846448-fwbxb" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.918697 4876 generic.go:334] "Generic (PLEG): container finished" podID="1e638147-a228-4f35-b3f1-8ead1cc5db48" containerID="117e5225f0ecf9e60d6a447160902698a1ba249404e7f25f760ef4f1bb4fe123" exitCode=0 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.918885 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9cljp" event={"ID":"1e638147-a228-4f35-b3f1-8ead1cc5db48","Type":"ContainerDied","Data":"117e5225f0ecf9e60d6a447160902698a1ba249404e7f25f760ef4f1bb4fe123"} Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.921406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerStarted","Data":"0aac21b5a4ebe56924dedfc9b14e37014510285557c08eb2c5cc09ef7266a937"} Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.921560 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="ceilometer-central-agent" containerID="cri-o://282186d79c12f24f859b943d17c1a42f3ec95ee5a69a9a7c212c1b6b129d6d66" gracePeriod=30 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.922761 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="proxy-httpd" containerID="cri-o://0aac21b5a4ebe56924dedfc9b14e37014510285557c08eb2c5cc09ef7266a937" gracePeriod=30 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.922961 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="sg-core" containerID="cri-o://a486f16070c5ab7dc259453d73a209740231be7646a624574b751659ed9b079c" gracePeriod=30 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.923013 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="ceilometer-notification-agent" containerID="cri-o://6d2cd6ea6447c668eb5814192932fc05d8c0025eae5bb579833725ff82bbecc7" gracePeriod=30 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.923162 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.926658 4876 generic.go:334] "Generic (PLEG): container finished" podID="ed3eafa5-4131-4235-aba0-de93198af37d" containerID="a591cb44f594d6588a92f9089754f675bf3be8a1ab2a2a8a17c47ffa58952fbd" exitCode=0 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.926725 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d661-account-create-update-cb7c4" event={"ID":"ed3eafa5-4131-4235-aba0-de93198af37d","Type":"ContainerDied","Data":"a591cb44f594d6588a92f9089754f675bf3be8a1ab2a2a8a17c47ffa58952fbd"} Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.941149 4876 generic.go:334] "Generic (PLEG): container finished" podID="d2bce654-ab4b-4687-8216-21761f83a696" containerID="16ca771999698bdf2bf33649c6cb9a6aa528f806ec5c1766673c609bc8d43e45" exitCode=0 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.941210 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" event={"ID":"d2bce654-ab4b-4687-8216-21761f83a696","Type":"ContainerDied","Data":"16ca771999698bdf2bf33649c6cb9a6aa528f806ec5c1766673c609bc8d43e45"} Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.943957 4876 generic.go:334] "Generic (PLEG): container finished" podID="f70aa4ff-ac52-4cef-808e-6133ef7d395f" containerID="5e0777b08a3d7a1ed3ab11d5a47a1bc7dbe631480df9c22897e474b120904b6f" exitCode=0 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.944009 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hb5wk" event={"ID":"f70aa4ff-ac52-4cef-808e-6133ef7d395f","Type":"ContainerDied","Data":"5e0777b08a3d7a1ed3ab11d5a47a1bc7dbe631480df9c22897e474b120904b6f"} Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.946458 4876 generic.go:334] "Generic (PLEG): container finished" podID="693e7fd3-fc4e-44fe-827c-2063e9d2f496" containerID="b405581d05fc048054062d7341eb0a64d71fe644de6c1d05cd459c513614b0ac" exitCode=0 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.946531 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" event={"ID":"693e7fd3-fc4e-44fe-827c-2063e9d2f496","Type":"ContainerDied","Data":"b405581d05fc048054062d7341eb0a64d71fe644de6c1d05cd459c513614b0ac"} Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.947956 4876 generic.go:334] "Generic (PLEG): container finished" podID="9b58b1a1-c690-4a1c-9280-09c5baa7a5a3" containerID="d2bffe9292c162769ac09a3e456a25f0559b47668f715d740b0bcedbd0ca85e9" exitCode=0 Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.948015 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z22v8" event={"ID":"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3","Type":"ContainerDied","Data":"d2bffe9292c162769ac09a3e456a25f0559b47668f715d740b0bcedbd0ca85e9"} Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.948835 4876 scope.go:117] "RemoveContainer" containerID="cb3715ab52291aa14e991b69ebb5e5514f25842602da4b54043af15ed7dd7401" Dec 05 08:37:14 crc kubenswrapper[4876]: E1205 08:37:14.949196 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-74677b69bc-66gfk_openstack(2066efee-0c4c-4e8d-913f-2cb7782b5863)\"" pod="openstack/heat-api-74677b69bc-66gfk" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" Dec 05 08:37:14 crc kubenswrapper[4876]: I1205 08:37:14.968414 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.364911665 podStartE2EDuration="6.96839337s" podCreationTimestamp="2025-12-05 08:37:08 +0000 UTC" firstStartedPulling="2025-12-05 08:37:09.886715143 +0000 UTC m=+1294.375379755" lastFinishedPulling="2025-12-05 08:37:14.490196838 +0000 UTC m=+1298.978861460" observedRunningTime="2025-12-05 08:37:14.958204558 +0000 UTC m=+1299.446869180" watchObservedRunningTime="2025-12-05 08:37:14.96839337 +0000 UTC m=+1299.457057982" Dec 05 08:37:15 crc kubenswrapper[4876]: I1205 08:37:15.613780 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:37:15 crc kubenswrapper[4876]: I1205 08:37:15.960609 4876 generic.go:334] "Generic (PLEG): container finished" podID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerID="0aac21b5a4ebe56924dedfc9b14e37014510285557c08eb2c5cc09ef7266a937" exitCode=0 Dec 05 08:37:15 crc kubenswrapper[4876]: I1205 08:37:15.960643 4876 generic.go:334] "Generic (PLEG): container finished" podID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerID="a486f16070c5ab7dc259453d73a209740231be7646a624574b751659ed9b079c" exitCode=2 Dec 05 08:37:15 crc kubenswrapper[4876]: I1205 08:37:15.960651 4876 generic.go:334] "Generic (PLEG): container finished" podID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerID="6d2cd6ea6447c668eb5814192932fc05d8c0025eae5bb579833725ff82bbecc7" exitCode=0 Dec 05 08:37:15 crc kubenswrapper[4876]: I1205 08:37:15.960741 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerDied","Data":"0aac21b5a4ebe56924dedfc9b14e37014510285557c08eb2c5cc09ef7266a937"} Dec 05 08:37:15 crc kubenswrapper[4876]: I1205 08:37:15.960794 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerDied","Data":"a486f16070c5ab7dc259453d73a209740231be7646a624574b751659ed9b079c"} Dec 05 08:37:15 crc kubenswrapper[4876]: I1205 08:37:15.960806 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerDied","Data":"6d2cd6ea6447c668eb5814192932fc05d8c0025eae5bb579833725ff82bbecc7"} Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.535956 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.661757 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fp98\" (UniqueName: \"kubernetes.io/projected/d2bce654-ab4b-4687-8216-21761f83a696-kube-api-access-6fp98\") pod \"d2bce654-ab4b-4687-8216-21761f83a696\" (UID: \"d2bce654-ab4b-4687-8216-21761f83a696\") " Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.681426 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bce654-ab4b-4687-8216-21761f83a696-operator-scripts\") pod \"d2bce654-ab4b-4687-8216-21761f83a696\" (UID: \"d2bce654-ab4b-4687-8216-21761f83a696\") " Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.675979 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2bce654-ab4b-4687-8216-21761f83a696-kube-api-access-6fp98" (OuterVolumeSpecName: "kube-api-access-6fp98") pod "d2bce654-ab4b-4687-8216-21761f83a696" (UID: "d2bce654-ab4b-4687-8216-21761f83a696"). InnerVolumeSpecName "kube-api-access-6fp98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.684369 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fp98\" (UniqueName: \"kubernetes.io/projected/d2bce654-ab4b-4687-8216-21761f83a696-kube-api-access-6fp98\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.685041 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2bce654-ab4b-4687-8216-21761f83a696-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2bce654-ab4b-4687-8216-21761f83a696" (UID: "d2bce654-ab4b-4687-8216-21761f83a696"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.787229 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2bce654-ab4b-4687-8216-21761f83a696-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.958261 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.965227 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.973213 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.975069 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z22v8" event={"ID":"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3","Type":"ContainerDied","Data":"52b3a4f3b2c53c6e2a6471a3ddb38355dd09647c35ad0df039ab173b7693271b"} Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.975210 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52b3a4f3b2c53c6e2a6471a3ddb38355dd09647c35ad0df039ab173b7693271b" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.975322 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z22v8" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.983328 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.986077 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9cljp" event={"ID":"1e638147-a228-4f35-b3f1-8ead1cc5db48","Type":"ContainerDied","Data":"071624cd9f8b7f6e72d55c4da060c4142f3567e5c53ef1a2adf768718fc312b0"} Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.986125 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="071624cd9f8b7f6e72d55c4da060c4142f3567e5c53ef1a2adf768718fc312b0" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.988461 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-d661-account-create-update-cb7c4" event={"ID":"ed3eafa5-4131-4235-aba0-de93198af37d","Type":"ContainerDied","Data":"9d97da2f6d71e512b99ff23d5d6b739f205d5c1ca6a6106ea2944ac4b6bbcd5f"} Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.988486 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d97da2f6d71e512b99ff23d5d6b739f205d5c1ca6a6106ea2944ac4b6bbcd5f" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.992015 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-d661-account-create-update-cb7c4" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.997939 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" event={"ID":"d2bce654-ab4b-4687-8216-21761f83a696","Type":"ContainerDied","Data":"ce4482fff8dcf003080d40a22b232ef255b6110489ab9a16c1283aabe4317117"} Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.998149 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce4482fff8dcf003080d40a22b232ef255b6110489ab9a16c1283aabe4317117" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.998273 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-87f7-account-create-update-fsgxj" Dec 05 08:37:16 crc kubenswrapper[4876]: I1205 08:37:16.999592 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.000690 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hb5wk" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.000771 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hb5wk" event={"ID":"f70aa4ff-ac52-4cef-808e-6133ef7d395f","Type":"ContainerDied","Data":"ee6a584c12138550fdb9b7e08cc4c6fe4521234f899036b256c27465d360c351"} Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.000812 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee6a584c12138550fdb9b7e08cc4c6fe4521234f899036b256c27465d360c351" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.002856 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" event={"ID":"693e7fd3-fc4e-44fe-827c-2063e9d2f496","Type":"ContainerDied","Data":"19c4e5715430a633b8c0be21e66e431c4ef682437d84041c63e2a6084a4106a4"} Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.002879 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19c4e5715430a633b8c0be21e66e431c4ef682437d84041c63e2a6084a4106a4" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.002928 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b05e-account-create-update-5t4rh" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092178 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-operator-scripts\") pod \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\" (UID: \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092244 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76pdw\" (UniqueName: \"kubernetes.io/projected/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-kube-api-access-76pdw\") pod \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\" (UID: \"9b58b1a1-c690-4a1c-9280-09c5baa7a5a3\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092269 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sw8j\" (UniqueName: \"kubernetes.io/projected/1e638147-a228-4f35-b3f1-8ead1cc5db48-kube-api-access-7sw8j\") pod \"1e638147-a228-4f35-b3f1-8ead1cc5db48\" (UID: \"1e638147-a228-4f35-b3f1-8ead1cc5db48\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092323 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e638147-a228-4f35-b3f1-8ead1cc5db48-operator-scripts\") pod \"1e638147-a228-4f35-b3f1-8ead1cc5db48\" (UID: \"1e638147-a228-4f35-b3f1-8ead1cc5db48\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092359 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh9x2\" (UniqueName: \"kubernetes.io/projected/ed3eafa5-4131-4235-aba0-de93198af37d-kube-api-access-xh9x2\") pod \"ed3eafa5-4131-4235-aba0-de93198af37d\" (UID: \"ed3eafa5-4131-4235-aba0-de93198af37d\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092408 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nccfg\" (UniqueName: \"kubernetes.io/projected/f70aa4ff-ac52-4cef-808e-6133ef7d395f-kube-api-access-nccfg\") pod \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\" (UID: \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092434 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3eafa5-4131-4235-aba0-de93198af37d-operator-scripts\") pod \"ed3eafa5-4131-4235-aba0-de93198af37d\" (UID: \"ed3eafa5-4131-4235-aba0-de93198af37d\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092464 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcfnt\" (UniqueName: \"kubernetes.io/projected/693e7fd3-fc4e-44fe-827c-2063e9d2f496-kube-api-access-fcfnt\") pod \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\" (UID: \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092506 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70aa4ff-ac52-4cef-808e-6133ef7d395f-operator-scripts\") pod \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\" (UID: \"f70aa4ff-ac52-4cef-808e-6133ef7d395f\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.092535 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/693e7fd3-fc4e-44fe-827c-2063e9d2f496-operator-scripts\") pod \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\" (UID: \"693e7fd3-fc4e-44fe-827c-2063e9d2f496\") " Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.093822 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693e7fd3-fc4e-44fe-827c-2063e9d2f496-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "693e7fd3-fc4e-44fe-827c-2063e9d2f496" (UID: "693e7fd3-fc4e-44fe-827c-2063e9d2f496"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.094133 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9b58b1a1-c690-4a1c-9280-09c5baa7a5a3" (UID: "9b58b1a1-c690-4a1c-9280-09c5baa7a5a3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.094501 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e638147-a228-4f35-b3f1-8ead1cc5db48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e638147-a228-4f35-b3f1-8ead1cc5db48" (UID: "1e638147-a228-4f35-b3f1-8ead1cc5db48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.096302 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f70aa4ff-ac52-4cef-808e-6133ef7d395f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f70aa4ff-ac52-4cef-808e-6133ef7d395f" (UID: "f70aa4ff-ac52-4cef-808e-6133ef7d395f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.096655 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed3eafa5-4131-4235-aba0-de93198af37d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed3eafa5-4131-4235-aba0-de93198af37d" (UID: "ed3eafa5-4131-4235-aba0-de93198af37d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.097806 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-kube-api-access-76pdw" (OuterVolumeSpecName: "kube-api-access-76pdw") pod "9b58b1a1-c690-4a1c-9280-09c5baa7a5a3" (UID: "9b58b1a1-c690-4a1c-9280-09c5baa7a5a3"). InnerVolumeSpecName "kube-api-access-76pdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.099105 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e638147-a228-4f35-b3f1-8ead1cc5db48-kube-api-access-7sw8j" (OuterVolumeSpecName: "kube-api-access-7sw8j") pod "1e638147-a228-4f35-b3f1-8ead1cc5db48" (UID: "1e638147-a228-4f35-b3f1-8ead1cc5db48"). InnerVolumeSpecName "kube-api-access-7sw8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.101581 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/693e7fd3-fc4e-44fe-827c-2063e9d2f496-kube-api-access-fcfnt" (OuterVolumeSpecName: "kube-api-access-fcfnt") pod "693e7fd3-fc4e-44fe-827c-2063e9d2f496" (UID: "693e7fd3-fc4e-44fe-827c-2063e9d2f496"). InnerVolumeSpecName "kube-api-access-fcfnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.106303 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed3eafa5-4131-4235-aba0-de93198af37d-kube-api-access-xh9x2" (OuterVolumeSpecName: "kube-api-access-xh9x2") pod "ed3eafa5-4131-4235-aba0-de93198af37d" (UID: "ed3eafa5-4131-4235-aba0-de93198af37d"). InnerVolumeSpecName "kube-api-access-xh9x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.108635 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f70aa4ff-ac52-4cef-808e-6133ef7d395f-kube-api-access-nccfg" (OuterVolumeSpecName: "kube-api-access-nccfg") pod "f70aa4ff-ac52-4cef-808e-6133ef7d395f" (UID: "f70aa4ff-ac52-4cef-808e-6133ef7d395f"). InnerVolumeSpecName "kube-api-access-nccfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.194468 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.194720 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76pdw\" (UniqueName: \"kubernetes.io/projected/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3-kube-api-access-76pdw\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.194783 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sw8j\" (UniqueName: \"kubernetes.io/projected/1e638147-a228-4f35-b3f1-8ead1cc5db48-kube-api-access-7sw8j\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.194850 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e638147-a228-4f35-b3f1-8ead1cc5db48-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.194925 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh9x2\" (UniqueName: \"kubernetes.io/projected/ed3eafa5-4131-4235-aba0-de93198af37d-kube-api-access-xh9x2\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.194988 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nccfg\" (UniqueName: \"kubernetes.io/projected/f70aa4ff-ac52-4cef-808e-6133ef7d395f-kube-api-access-nccfg\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.195057 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed3eafa5-4131-4235-aba0-de93198af37d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.195111 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcfnt\" (UniqueName: \"kubernetes.io/projected/693e7fd3-fc4e-44fe-827c-2063e9d2f496-kube-api-access-fcfnt\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.195183 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f70aa4ff-ac52-4cef-808e-6133ef7d395f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.195246 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/693e7fd3-fc4e-44fe-827c-2063e9d2f496-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.765376 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-76699577f6-nt9wb" Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.841127 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-767d846448-fwbxb"] Dec 05 08:37:17 crc kubenswrapper[4876]: I1205 08:37:17.998882 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5f87bc649b-spcsf" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.012431 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9cljp" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.091284 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-74677b69bc-66gfk"] Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.379026 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.492139 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.528144 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s9rd\" (UniqueName: \"kubernetes.io/projected/6c7d34c2-eacd-4cf3-840b-08ab27365769-kube-api-access-2s9rd\") pod \"6c7d34c2-eacd-4cf3-840b-08ab27365769\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.528335 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-combined-ca-bundle\") pod \"6c7d34c2-eacd-4cf3-840b-08ab27365769\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.528462 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data\") pod \"6c7d34c2-eacd-4cf3-840b-08ab27365769\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.528497 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data-custom\") pod \"6c7d34c2-eacd-4cf3-840b-08ab27365769\" (UID: \"6c7d34c2-eacd-4cf3-840b-08ab27365769\") " Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.543738 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6c7d34c2-eacd-4cf3-840b-08ab27365769" (UID: "6c7d34c2-eacd-4cf3-840b-08ab27365769"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.550409 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c7d34c2-eacd-4cf3-840b-08ab27365769-kube-api-access-2s9rd" (OuterVolumeSpecName: "kube-api-access-2s9rd") pod "6c7d34c2-eacd-4cf3-840b-08ab27365769" (UID: "6c7d34c2-eacd-4cf3-840b-08ab27365769"). InnerVolumeSpecName "kube-api-access-2s9rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.570889 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c7d34c2-eacd-4cf3-840b-08ab27365769" (UID: "6c7d34c2-eacd-4cf3-840b-08ab27365769"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.580226 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data" (OuterVolumeSpecName: "config-data") pod "6c7d34c2-eacd-4cf3-840b-08ab27365769" (UID: "6c7d34c2-eacd-4cf3-840b-08ab27365769"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.622865 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-98b56b8f5-ksg4z" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.631130 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6brj\" (UniqueName: \"kubernetes.io/projected/2066efee-0c4c-4e8d-913f-2cb7782b5863-kube-api-access-h6brj\") pod \"2066efee-0c4c-4e8d-913f-2cb7782b5863\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.631201 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-combined-ca-bundle\") pod \"2066efee-0c4c-4e8d-913f-2cb7782b5863\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.631277 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data\") pod \"2066efee-0c4c-4e8d-913f-2cb7782b5863\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.631465 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data-custom\") pod \"2066efee-0c4c-4e8d-913f-2cb7782b5863\" (UID: \"2066efee-0c4c-4e8d-913f-2cb7782b5863\") " Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.631971 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s9rd\" (UniqueName: \"kubernetes.io/projected/6c7d34c2-eacd-4cf3-840b-08ab27365769-kube-api-access-2s9rd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.631995 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.632007 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.632020 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c7d34c2-eacd-4cf3-840b-08ab27365769-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.634560 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2066efee-0c4c-4e8d-913f-2cb7782b5863-kube-api-access-h6brj" (OuterVolumeSpecName: "kube-api-access-h6brj") pod "2066efee-0c4c-4e8d-913f-2cb7782b5863" (UID: "2066efee-0c4c-4e8d-913f-2cb7782b5863"). InnerVolumeSpecName "kube-api-access-h6brj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.636347 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2066efee-0c4c-4e8d-913f-2cb7782b5863" (UID: "2066efee-0c4c-4e8d-913f-2cb7782b5863"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.669767 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2066efee-0c4c-4e8d-913f-2cb7782b5863" (UID: "2066efee-0c4c-4e8d-913f-2cb7782b5863"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.714986 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data" (OuterVolumeSpecName: "config-data") pod "2066efee-0c4c-4e8d-913f-2cb7782b5863" (UID: "2066efee-0c4c-4e8d-913f-2cb7782b5863"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.733871 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.733990 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6brj\" (UniqueName: \"kubernetes.io/projected/2066efee-0c4c-4e8d-913f-2cb7782b5863-kube-api-access-h6brj\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.734000 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:18 crc kubenswrapper[4876]: I1205 08:37:18.734009 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066efee-0c4c-4e8d-913f-2cb7782b5863-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.022017 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-74677b69bc-66gfk" Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.022040 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-74677b69bc-66gfk" event={"ID":"2066efee-0c4c-4e8d-913f-2cb7782b5863","Type":"ContainerDied","Data":"af0445a57870ffa94bf5a3251a1e6cab22c8a01ca5d944b3d04c610b4433b596"} Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.022484 4876 scope.go:117] "RemoveContainer" containerID="cb3715ab52291aa14e991b69ebb5e5514f25842602da4b54043af15ed7dd7401" Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.030800 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-767d846448-fwbxb" event={"ID":"6c7d34c2-eacd-4cf3-840b-08ab27365769","Type":"ContainerDied","Data":"2444ae922e59f4a83523fec1f189d2b24386c4571d2ba2f3507046877ee28c40"} Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.030839 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-767d846448-fwbxb" Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.061764 4876 scope.go:117] "RemoveContainer" containerID="d92f1b42a0ff4c27934987cf8c1c33b97a985461e67bc370a23a26cc43f61e32" Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.072211 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-74677b69bc-66gfk"] Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.084914 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-74677b69bc-66gfk"] Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.094475 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-767d846448-fwbxb"] Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.106307 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-767d846448-fwbxb"] Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.835093 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" path="/var/lib/kubelet/pods/2066efee-0c4c-4e8d-913f-2cb7782b5863/volumes" Dec 05 08:37:19 crc kubenswrapper[4876]: I1205 08:37:19.835675 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" path="/var/lib/kubelet/pods/6c7d34c2-eacd-4cf3-840b-08ab27365769/volumes" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.180668 4876 generic.go:334] "Generic (PLEG): container finished" podID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerID="282186d79c12f24f859b943d17c1a42f3ec95ee5a69a9a7c212c1b6b129d6d66" exitCode=0 Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.180700 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerDied","Data":"282186d79c12f24f859b943d17c1a42f3ec95ee5a69a9a7c212c1b6b129d6d66"} Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.278439 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.302675 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-run-httpd\") pod \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.302821 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-log-httpd\") pod \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.303086 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-config-data\") pod \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.303126 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "03bf85da-ca2e-4186-96cb-8a2f0373b2b3" (UID: "03bf85da-ca2e-4186-96cb-8a2f0373b2b3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.303128 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-combined-ca-bundle\") pod \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.303193 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-scripts\") pod \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.303252 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9zvv\" (UniqueName: \"kubernetes.io/projected/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-kube-api-access-f9zvv\") pod \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.303313 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "03bf85da-ca2e-4186-96cb-8a2f0373b2b3" (UID: "03bf85da-ca2e-4186-96cb-8a2f0373b2b3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.303351 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-sg-core-conf-yaml\") pod \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\" (UID: \"03bf85da-ca2e-4186-96cb-8a2f0373b2b3\") " Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.304037 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.304086 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.322117 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-kube-api-access-f9zvv" (OuterVolumeSpecName: "kube-api-access-f9zvv") pod "03bf85da-ca2e-4186-96cb-8a2f0373b2b3" (UID: "03bf85da-ca2e-4186-96cb-8a2f0373b2b3"). InnerVolumeSpecName "kube-api-access-f9zvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.341138 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-scripts" (OuterVolumeSpecName: "scripts") pod "03bf85da-ca2e-4186-96cb-8a2f0373b2b3" (UID: "03bf85da-ca2e-4186-96cb-8a2f0373b2b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.406244 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.406281 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9zvv\" (UniqueName: \"kubernetes.io/projected/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-kube-api-access-f9zvv\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.427227 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "03bf85da-ca2e-4186-96cb-8a2f0373b2b3" (UID: "03bf85da-ca2e-4186-96cb-8a2f0373b2b3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.466096 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03bf85da-ca2e-4186-96cb-8a2f0373b2b3" (UID: "03bf85da-ca2e-4186-96cb-8a2f0373b2b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.477063 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-config-data" (OuterVolumeSpecName: "config-data") pod "03bf85da-ca2e-4186-96cb-8a2f0373b2b3" (UID: "03bf85da-ca2e-4186-96cb-8a2f0373b2b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.508144 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.508177 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:22 crc kubenswrapper[4876]: I1205 08:37:22.508190 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03bf85da-ca2e-4186-96cb-8a2f0373b2b3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.191208 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03bf85da-ca2e-4186-96cb-8a2f0373b2b3","Type":"ContainerDied","Data":"c9b24e08a18a0afd5a75242486bcc0a8670b99dcb5b9af75d5684350c785013f"} Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.191262 4876 scope.go:117] "RemoveContainer" containerID="0aac21b5a4ebe56924dedfc9b14e37014510285557c08eb2c5cc09ef7266a937" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.191300 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.219407 4876 scope.go:117] "RemoveContainer" containerID="a486f16070c5ab7dc259453d73a209740231be7646a624574b751659ed9b079c" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221032 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-54h52"] Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221499 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70aa4ff-ac52-4cef-808e-6133ef7d395f" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221514 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70aa4ff-ac52-4cef-808e-6133ef7d395f" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221543 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693e7fd3-fc4e-44fe-827c-2063e9d2f496" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221550 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="693e7fd3-fc4e-44fe-827c-2063e9d2f496" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221561 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" containerName="heat-api" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221567 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" containerName="heat-api" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221579 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="sg-core" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221585 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="sg-core" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221592 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" containerName="heat-api" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221639 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" containerName="heat-api" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221651 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" containerName="heat-cfnapi" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221657 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" containerName="heat-cfnapi" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221666 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="ceilometer-central-agent" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221673 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="ceilometer-central-agent" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221730 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2bce654-ab4b-4687-8216-21761f83a696" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221736 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2bce654-ab4b-4687-8216-21761f83a696" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221748 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b58b1a1-c690-4a1c-9280-09c5baa7a5a3" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221754 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b58b1a1-c690-4a1c-9280-09c5baa7a5a3" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221764 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="ceilometer-notification-agent" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221786 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="ceilometer-notification-agent" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221795 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e638147-a228-4f35-b3f1-8ead1cc5db48" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221800 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e638147-a228-4f35-b3f1-8ead1cc5db48" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221810 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="proxy-httpd" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221816 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="proxy-httpd" Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.221825 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed3eafa5-4131-4235-aba0-de93198af37d" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.221831 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed3eafa5-4131-4235-aba0-de93198af37d" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222040 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="proxy-httpd" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222067 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" containerName="heat-cfnapi" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222075 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e638147-a228-4f35-b3f1-8ead1cc5db48" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222085 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" containerName="heat-cfnapi" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222095 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed3eafa5-4131-4235-aba0-de93198af37d" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222104 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="ceilometer-central-agent" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222118 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" containerName="heat-api" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222144 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f70aa4ff-ac52-4cef-808e-6133ef7d395f" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222153 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b58b1a1-c690-4a1c-9280-09c5baa7a5a3" containerName="mariadb-database-create" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222162 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2bce654-ab4b-4687-8216-21761f83a696" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222168 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="693e7fd3-fc4e-44fe-827c-2063e9d2f496" containerName="mariadb-account-create-update" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222177 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="sg-core" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222186 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" containerName="ceilometer-notification-agent" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.222916 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.225975 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.226540 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.226840 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-x7tb8" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.240395 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.245558 4876 scope.go:117] "RemoveContainer" containerID="6d2cd6ea6447c668eb5814192932fc05d8c0025eae5bb579833725ff82bbecc7" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.256175 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.265347 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-54h52"] Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.287263 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:23 crc kubenswrapper[4876]: E1205 08:37:23.287711 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" containerName="heat-cfnapi" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.287724 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7d34c2-eacd-4cf3-840b-08ab27365769" containerName="heat-cfnapi" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.287954 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2066efee-0c4c-4e8d-913f-2cb7782b5863" containerName="heat-api" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.289534 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.292054 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.292277 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.296259 4876 scope.go:117] "RemoveContainer" containerID="282186d79c12f24f859b943d17c1a42f3ec95ee5a69a9a7c212c1b6b129d6d66" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.297741 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.322625 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-log-httpd\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.322714 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-scripts\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.322774 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.322804 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt2k7\" (UniqueName: \"kubernetes.io/projected/54116403-e4a3-4fb8-a94a-09d43e9f5972-kube-api-access-kt2k7\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.322820 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-run-httpd\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.322852 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-scripts\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.322869 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbkxt\" (UniqueName: \"kubernetes.io/projected/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-kube-api-access-wbkxt\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.323022 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.323049 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.323105 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-config-data\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.323161 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-config-data\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.424792 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-log-httpd\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.424891 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-scripts\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.424964 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425033 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt2k7\" (UniqueName: \"kubernetes.io/projected/54116403-e4a3-4fb8-a94a-09d43e9f5972-kube-api-access-kt2k7\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425063 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-run-httpd\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425087 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-scripts\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425113 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbkxt\" (UniqueName: \"kubernetes.io/projected/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-kube-api-access-wbkxt\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425169 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425200 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425240 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-config-data\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425297 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-config-data\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425960 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-log-httpd\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.425987 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-run-httpd\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.431581 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.432577 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.433712 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-scripts\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.433965 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.434511 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-config-data\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.440556 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-scripts\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.442668 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt2k7\" (UniqueName: \"kubernetes.io/projected/54116403-e4a3-4fb8-a94a-09d43e9f5972-kube-api-access-kt2k7\") pod \"ceilometer-0\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.443008 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-config-data\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.444370 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbkxt\" (UniqueName: \"kubernetes.io/projected/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-kube-api-access-wbkxt\") pod \"nova-cell0-conductor-db-sync-54h52\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.549292 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.614229 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:23 crc kubenswrapper[4876]: I1205 08:37:23.833424 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03bf85da-ca2e-4186-96cb-8a2f0373b2b3" path="/var/lib/kubelet/pods/03bf85da-ca2e-4186-96cb-8a2f0373b2b3/volumes" Dec 05 08:37:24 crc kubenswrapper[4876]: I1205 08:37:24.016409 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-54h52"] Dec 05 08:37:24 crc kubenswrapper[4876]: W1205 08:37:24.021127 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99d7f1cf_79c4_42c4_bd75_a6588e8daaec.slice/crio-74fe2a9c92335a45e2653074cc18963f99ac6d873205236a31e858f1e3a2f583 WatchSource:0}: Error finding container 74fe2a9c92335a45e2653074cc18963f99ac6d873205236a31e858f1e3a2f583: Status 404 returned error can't find the container with id 74fe2a9c92335a45e2653074cc18963f99ac6d873205236a31e858f1e3a2f583 Dec 05 08:37:24 crc kubenswrapper[4876]: I1205 08:37:24.169569 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:24 crc kubenswrapper[4876]: W1205 08:37:24.174090 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54116403_e4a3_4fb8_a94a_09d43e9f5972.slice/crio-ae47d5b62ff4ee6e6a6386d76fe57cb8fe2110e72b33caf2ed6a399a7b880a57 WatchSource:0}: Error finding container ae47d5b62ff4ee6e6a6386d76fe57cb8fe2110e72b33caf2ed6a399a7b880a57: Status 404 returned error can't find the container with id ae47d5b62ff4ee6e6a6386d76fe57cb8fe2110e72b33caf2ed6a399a7b880a57 Dec 05 08:37:24 crc kubenswrapper[4876]: I1205 08:37:24.203448 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-54h52" event={"ID":"99d7f1cf-79c4-42c4-bd75-a6588e8daaec","Type":"ContainerStarted","Data":"74fe2a9c92335a45e2653074cc18963f99ac6d873205236a31e858f1e3a2f583"} Dec 05 08:37:24 crc kubenswrapper[4876]: I1205 08:37:24.204949 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerStarted","Data":"ae47d5b62ff4ee6e6a6386d76fe57cb8fe2110e72b33caf2ed6a399a7b880a57"} Dec 05 08:37:24 crc kubenswrapper[4876]: I1205 08:37:24.393489 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-b64b8b465-trj6k" Dec 05 08:37:24 crc kubenswrapper[4876]: I1205 08:37:24.447022 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-689f7d9f48-fjhvt"] Dec 05 08:37:24 crc kubenswrapper[4876]: I1205 08:37:24.449607 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-689f7d9f48-fjhvt" podUID="8b5c7095-36f0-4024-b430-714d5603e08b" containerName="heat-engine" containerID="cri-o://340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" gracePeriod=60 Dec 05 08:37:25 crc kubenswrapper[4876]: I1205 08:37:25.216380 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerStarted","Data":"5217fce21e23ae62d8c63aa268f34b3b69e3bac1750450dd6284ed6c9b7f29a4"} Dec 05 08:37:25 crc kubenswrapper[4876]: E1205 08:37:25.592757 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 05 08:37:25 crc kubenswrapper[4876]: E1205 08:37:25.612024 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 05 08:37:25 crc kubenswrapper[4876]: E1205 08:37:25.619184 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 05 08:37:25 crc kubenswrapper[4876]: E1205 08:37:25.619253 4876 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-689f7d9f48-fjhvt" podUID="8b5c7095-36f0-4024-b430-714d5603e08b" containerName="heat-engine" Dec 05 08:37:26 crc kubenswrapper[4876]: I1205 08:37:26.227506 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerStarted","Data":"fdc56d6d3839f8fa067d6131fcaf7dacc424c89020d0c97ebe6ba27e4d052230"} Dec 05 08:37:27 crc kubenswrapper[4876]: I1205 08:37:27.247988 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerStarted","Data":"1a906aeac5f8f51fd0f2289b6cab9c070a2c97b8d667b5403209c5fe89b95855"} Dec 05 08:37:29 crc kubenswrapper[4876]: I1205 08:37:29.026295 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:35 crc kubenswrapper[4876]: E1205 08:37:35.586364 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 05 08:37:35 crc kubenswrapper[4876]: E1205 08:37:35.588867 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 05 08:37:35 crc kubenswrapper[4876]: E1205 08:37:35.590870 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 05 08:37:35 crc kubenswrapper[4876]: E1205 08:37:35.590915 4876 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-689f7d9f48-fjhvt" podUID="8b5c7095-36f0-4024-b430-714d5603e08b" containerName="heat-engine" Dec 05 08:37:36 crc kubenswrapper[4876]: I1205 08:37:36.349258 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerStarted","Data":"c4ad1274db568b32411facd8f45102aa901539fc8857ba53b09ea0a8b5768ed2"} Dec 05 08:37:36 crc kubenswrapper[4876]: I1205 08:37:36.349515 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="proxy-httpd" containerID="cri-o://c4ad1274db568b32411facd8f45102aa901539fc8857ba53b09ea0a8b5768ed2" gracePeriod=30 Dec 05 08:37:36 crc kubenswrapper[4876]: I1205 08:37:36.349558 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="ceilometer-central-agent" containerID="cri-o://5217fce21e23ae62d8c63aa268f34b3b69e3bac1750450dd6284ed6c9b7f29a4" gracePeriod=30 Dec 05 08:37:36 crc kubenswrapper[4876]: I1205 08:37:36.349555 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:37:36 crc kubenswrapper[4876]: I1205 08:37:36.349558 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="ceilometer-notification-agent" containerID="cri-o://fdc56d6d3839f8fa067d6131fcaf7dacc424c89020d0c97ebe6ba27e4d052230" gracePeriod=30 Dec 05 08:37:36 crc kubenswrapper[4876]: I1205 08:37:36.349508 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="sg-core" containerID="cri-o://1a906aeac5f8f51fd0f2289b6cab9c070a2c97b8d667b5403209c5fe89b95855" gracePeriod=30 Dec 05 08:37:36 crc kubenswrapper[4876]: I1205 08:37:36.379856 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=9.309656394 podStartE2EDuration="13.3798302s" podCreationTimestamp="2025-12-05 08:37:23 +0000 UTC" firstStartedPulling="2025-12-05 08:37:24.178233457 +0000 UTC m=+1308.666898079" lastFinishedPulling="2025-12-05 08:37:28.248407263 +0000 UTC m=+1312.737071885" observedRunningTime="2025-12-05 08:37:36.368024423 +0000 UTC m=+1320.856689055" watchObservedRunningTime="2025-12-05 08:37:36.3798302 +0000 UTC m=+1320.868494832" Dec 05 08:37:37 crc kubenswrapper[4876]: I1205 08:37:37.359455 4876 generic.go:334] "Generic (PLEG): container finished" podID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerID="c4ad1274db568b32411facd8f45102aa901539fc8857ba53b09ea0a8b5768ed2" exitCode=0 Dec 05 08:37:37 crc kubenswrapper[4876]: I1205 08:37:37.360019 4876 generic.go:334] "Generic (PLEG): container finished" podID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerID="1a906aeac5f8f51fd0f2289b6cab9c070a2c97b8d667b5403209c5fe89b95855" exitCode=2 Dec 05 08:37:37 crc kubenswrapper[4876]: I1205 08:37:37.360031 4876 generic.go:334] "Generic (PLEG): container finished" podID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerID="fdc56d6d3839f8fa067d6131fcaf7dacc424c89020d0c97ebe6ba27e4d052230" exitCode=0 Dec 05 08:37:37 crc kubenswrapper[4876]: I1205 08:37:37.359519 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerDied","Data":"c4ad1274db568b32411facd8f45102aa901539fc8857ba53b09ea0a8b5768ed2"} Dec 05 08:37:37 crc kubenswrapper[4876]: I1205 08:37:37.360387 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerDied","Data":"1a906aeac5f8f51fd0f2289b6cab9c070a2c97b8d667b5403209c5fe89b95855"} Dec 05 08:37:37 crc kubenswrapper[4876]: I1205 08:37:37.360412 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerDied","Data":"fdc56d6d3839f8fa067d6131fcaf7dacc424c89020d0c97ebe6ba27e4d052230"} Dec 05 08:37:37 crc kubenswrapper[4876]: I1205 08:37:37.361807 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-54h52" event={"ID":"99d7f1cf-79c4-42c4-bd75-a6588e8daaec","Type":"ContainerStarted","Data":"f87765f45920f1ea4498d1234b240f24872ab33ead370a56069f6fe647c0531e"} Dec 05 08:37:37 crc kubenswrapper[4876]: I1205 08:37:37.378807 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-54h52" podStartSLOduration=1.547025026 podStartE2EDuration="14.378790847s" podCreationTimestamp="2025-12-05 08:37:23 +0000 UTC" firstStartedPulling="2025-12-05 08:37:24.02346106 +0000 UTC m=+1308.512125692" lastFinishedPulling="2025-12-05 08:37:36.855226891 +0000 UTC m=+1321.343891513" observedRunningTime="2025-12-05 08:37:37.378206281 +0000 UTC m=+1321.866870903" watchObservedRunningTime="2025-12-05 08:37:37.378790847 +0000 UTC m=+1321.867455469" Dec 05 08:37:38 crc kubenswrapper[4876]: I1205 08:37:38.214166 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:37:38 crc kubenswrapper[4876]: I1205 08:37:38.214228 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:37:38 crc kubenswrapper[4876]: I1205 08:37:38.214272 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:37:38 crc kubenswrapper[4876]: I1205 08:37:38.215031 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9aa43e6caeca2c89f47c0cead848074edb893e13fe607a55570e6f144dca2a07"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:37:38 crc kubenswrapper[4876]: I1205 08:37:38.215092 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://9aa43e6caeca2c89f47c0cead848074edb893e13fe607a55570e6f144dca2a07" gracePeriod=600 Dec 05 08:37:38 crc kubenswrapper[4876]: I1205 08:37:38.378045 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="9aa43e6caeca2c89f47c0cead848074edb893e13fe607a55570e6f144dca2a07" exitCode=0 Dec 05 08:37:38 crc kubenswrapper[4876]: I1205 08:37:38.378130 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"9aa43e6caeca2c89f47c0cead848074edb893e13fe607a55570e6f144dca2a07"} Dec 05 08:37:38 crc kubenswrapper[4876]: I1205 08:37:38.378182 4876 scope.go:117] "RemoveContainer" containerID="6167cbd18c38bbbdb4c177b56aa429711e7d69fa0ae0c347c0b78af72fe60dcb" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.393509 4876 generic.go:334] "Generic (PLEG): container finished" podID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerID="5217fce21e23ae62d8c63aa268f34b3b69e3bac1750450dd6284ed6c9b7f29a4" exitCode=0 Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.393967 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerDied","Data":"5217fce21e23ae62d8c63aa268f34b3b69e3bac1750450dd6284ed6c9b7f29a4"} Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.396526 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8"} Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.650465 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.814884 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-scripts\") pod \"54116403-e4a3-4fb8-a94a-09d43e9f5972\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.815194 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-combined-ca-bundle\") pod \"54116403-e4a3-4fb8-a94a-09d43e9f5972\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.815230 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-run-httpd\") pod \"54116403-e4a3-4fb8-a94a-09d43e9f5972\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.815356 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-config-data\") pod \"54116403-e4a3-4fb8-a94a-09d43e9f5972\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.815603 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "54116403-e4a3-4fb8-a94a-09d43e9f5972" (UID: "54116403-e4a3-4fb8-a94a-09d43e9f5972"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.815765 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-sg-core-conf-yaml\") pod \"54116403-e4a3-4fb8-a94a-09d43e9f5972\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.815798 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt2k7\" (UniqueName: \"kubernetes.io/projected/54116403-e4a3-4fb8-a94a-09d43e9f5972-kube-api-access-kt2k7\") pod \"54116403-e4a3-4fb8-a94a-09d43e9f5972\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.816262 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-log-httpd\") pod \"54116403-e4a3-4fb8-a94a-09d43e9f5972\" (UID: \"54116403-e4a3-4fb8-a94a-09d43e9f5972\") " Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.816722 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "54116403-e4a3-4fb8-a94a-09d43e9f5972" (UID: "54116403-e4a3-4fb8-a94a-09d43e9f5972"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.817169 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.817248 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54116403-e4a3-4fb8-a94a-09d43e9f5972-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.833402 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54116403-e4a3-4fb8-a94a-09d43e9f5972-kube-api-access-kt2k7" (OuterVolumeSpecName: "kube-api-access-kt2k7") pod "54116403-e4a3-4fb8-a94a-09d43e9f5972" (UID: "54116403-e4a3-4fb8-a94a-09d43e9f5972"). InnerVolumeSpecName "kube-api-access-kt2k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.843496 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "54116403-e4a3-4fb8-a94a-09d43e9f5972" (UID: "54116403-e4a3-4fb8-a94a-09d43e9f5972"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.846453 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-scripts" (OuterVolumeSpecName: "scripts") pod "54116403-e4a3-4fb8-a94a-09d43e9f5972" (UID: "54116403-e4a3-4fb8-a94a-09d43e9f5972"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.918472 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.918499 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt2k7\" (UniqueName: \"kubernetes.io/projected/54116403-e4a3-4fb8-a94a-09d43e9f5972-kube-api-access-kt2k7\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.918510 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.934147 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54116403-e4a3-4fb8-a94a-09d43e9f5972" (UID: "54116403-e4a3-4fb8-a94a-09d43e9f5972"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:39 crc kubenswrapper[4876]: I1205 08:37:39.952425 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-config-data" (OuterVolumeSpecName: "config-data") pod "54116403-e4a3-4fb8-a94a-09d43e9f5972" (UID: "54116403-e4a3-4fb8-a94a-09d43e9f5972"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.019831 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.019864 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54116403-e4a3-4fb8-a94a-09d43e9f5972-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.155778 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.375434 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp5z9\" (UniqueName: \"kubernetes.io/projected/8b5c7095-36f0-4024-b430-714d5603e08b-kube-api-access-kp5z9\") pod \"8b5c7095-36f0-4024-b430-714d5603e08b\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.375841 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data\") pod \"8b5c7095-36f0-4024-b430-714d5603e08b\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.375959 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data-custom\") pod \"8b5c7095-36f0-4024-b430-714d5603e08b\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.376038 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-combined-ca-bundle\") pod \"8b5c7095-36f0-4024-b430-714d5603e08b\" (UID: \"8b5c7095-36f0-4024-b430-714d5603e08b\") " Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.380868 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8b5c7095-36f0-4024-b430-714d5603e08b" (UID: "8b5c7095-36f0-4024-b430-714d5603e08b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.381874 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b5c7095-36f0-4024-b430-714d5603e08b-kube-api-access-kp5z9" (OuterVolumeSpecName: "kube-api-access-kp5z9") pod "8b5c7095-36f0-4024-b430-714d5603e08b" (UID: "8b5c7095-36f0-4024-b430-714d5603e08b"). InnerVolumeSpecName "kube-api-access-kp5z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.410472 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.410470 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54116403-e4a3-4fb8-a94a-09d43e9f5972","Type":"ContainerDied","Data":"ae47d5b62ff4ee6e6a6386d76fe57cb8fe2110e72b33caf2ed6a399a7b880a57"} Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.410648 4876 scope.go:117] "RemoveContainer" containerID="c4ad1274db568b32411facd8f45102aa901539fc8857ba53b09ea0a8b5768ed2" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.412215 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b5c7095-36f0-4024-b430-714d5603e08b" (UID: "8b5c7095-36f0-4024-b430-714d5603e08b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.413262 4876 generic.go:334] "Generic (PLEG): container finished" podID="8b5c7095-36f0-4024-b430-714d5603e08b" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" exitCode=0 Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.413323 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-689f7d9f48-fjhvt" event={"ID":"8b5c7095-36f0-4024-b430-714d5603e08b","Type":"ContainerDied","Data":"340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722"} Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.413385 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-689f7d9f48-fjhvt" event={"ID":"8b5c7095-36f0-4024-b430-714d5603e08b","Type":"ContainerDied","Data":"fbb76d64ed72e6febfa177079798497bfecf76631dd7b9eb3c8c6767bc6a5de3"} Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.415263 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-689f7d9f48-fjhvt" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.435807 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data" (OuterVolumeSpecName: "config-data") pod "8b5c7095-36f0-4024-b430-714d5603e08b" (UID: "8b5c7095-36f0-4024-b430-714d5603e08b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.439577 4876 scope.go:117] "RemoveContainer" containerID="1a906aeac5f8f51fd0f2289b6cab9c070a2c97b8d667b5403209c5fe89b95855" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.454028 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.463599 4876 scope.go:117] "RemoveContainer" containerID="fdc56d6d3839f8fa067d6131fcaf7dacc424c89020d0c97ebe6ba27e4d052230" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.479975 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp5z9\" (UniqueName: \"kubernetes.io/projected/8b5c7095-36f0-4024-b430-714d5603e08b-kube-api-access-kp5z9\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.480015 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.480032 4876 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.480045 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b5c7095-36f0-4024-b430-714d5603e08b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.480645 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.497172 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:40 crc kubenswrapper[4876]: E1205 08:37:40.497592 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="ceilometer-notification-agent" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.497610 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="ceilometer-notification-agent" Dec 05 08:37:40 crc kubenswrapper[4876]: E1205 08:37:40.497681 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b5c7095-36f0-4024-b430-714d5603e08b" containerName="heat-engine" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.497690 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b5c7095-36f0-4024-b430-714d5603e08b" containerName="heat-engine" Dec 05 08:37:40 crc kubenswrapper[4876]: E1205 08:37:40.497738 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="sg-core" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.497748 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="sg-core" Dec 05 08:37:40 crc kubenswrapper[4876]: E1205 08:37:40.497760 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="proxy-httpd" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.497768 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="proxy-httpd" Dec 05 08:37:40 crc kubenswrapper[4876]: E1205 08:37:40.497793 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="ceilometer-central-agent" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.497803 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="ceilometer-central-agent" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.498073 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="ceilometer-central-agent" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.498090 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="proxy-httpd" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.498119 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="ceilometer-notification-agent" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.498127 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" containerName="sg-core" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.498151 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b5c7095-36f0-4024-b430-714d5603e08b" containerName="heat-engine" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.499573 4876 scope.go:117] "RemoveContainer" containerID="5217fce21e23ae62d8c63aa268f34b3b69e3bac1750450dd6284ed6c9b7f29a4" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.506918 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.508747 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.524265 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.524391 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.538615 4876 scope.go:117] "RemoveContainer" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.593359 4876 scope.go:117] "RemoveContainer" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" Dec 05 08:37:40 crc kubenswrapper[4876]: E1205 08:37:40.594477 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722\": container with ID starting with 340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722 not found: ID does not exist" containerID="340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.594520 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722"} err="failed to get container status \"340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722\": rpc error: code = NotFound desc = could not find container \"340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722\": container with ID starting with 340a69edb621e1e0e345e09ce969a79f222cc53286a6a530ca1602e524cbc722 not found: ID does not exist" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.684571 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.684842 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-run-httpd\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.684980 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9djcj\" (UniqueName: \"kubernetes.io/projected/aa99f31a-f42e-43c0-b104-b484332aa37e-kube-api-access-9djcj\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.685022 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.685073 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-log-httpd\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.685160 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-scripts\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.685191 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-config-data\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.812198 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.812276 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-log-httpd\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.812318 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-scripts\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.812349 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-config-data\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.812374 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.812472 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-run-httpd\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.812515 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9djcj\" (UniqueName: \"kubernetes.io/projected/aa99f31a-f42e-43c0-b104-b484332aa37e-kube-api-access-9djcj\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.813268 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-log-httpd\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.817118 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-run-httpd\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.819501 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.824198 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.824489 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-scripts\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.838962 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-config-data\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.844971 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9djcj\" (UniqueName: \"kubernetes.io/projected/aa99f31a-f42e-43c0-b104-b484332aa37e-kube-api-access-9djcj\") pod \"ceilometer-0\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " pod="openstack/ceilometer-0" Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.847697 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-689f7d9f48-fjhvt"] Dec 05 08:37:40 crc kubenswrapper[4876]: I1205 08:37:40.861284 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-689f7d9f48-fjhvt"] Dec 05 08:37:41 crc kubenswrapper[4876]: I1205 08:37:41.139817 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:41 crc kubenswrapper[4876]: I1205 08:37:41.837415 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54116403-e4a3-4fb8-a94a-09d43e9f5972" path="/var/lib/kubelet/pods/54116403-e4a3-4fb8-a94a-09d43e9f5972/volumes" Dec 05 08:37:41 crc kubenswrapper[4876]: I1205 08:37:41.839491 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b5c7095-36f0-4024-b430-714d5603e08b" path="/var/lib/kubelet/pods/8b5c7095-36f0-4024-b430-714d5603e08b/volumes" Dec 05 08:37:41 crc kubenswrapper[4876]: I1205 08:37:41.894828 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:42 crc kubenswrapper[4876]: I1205 08:37:42.465038 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerStarted","Data":"9f2eff702d021d79aea80826e0dc7eb8b92562790db99a759f03c86e1e1d76c3"} Dec 05 08:37:43 crc kubenswrapper[4876]: I1205 08:37:43.480954 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerStarted","Data":"2174f55d3c8fb34d9ec8b026c7acacf021fe0b3575d775c75c5bff54a352eff9"} Dec 05 08:37:44 crc kubenswrapper[4876]: I1205 08:37:44.490784 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerStarted","Data":"6f33d42dce2006d056f4f4006958d70a629e5a33a5d4ef461913f97da2de3b03"} Dec 05 08:37:44 crc kubenswrapper[4876]: I1205 08:37:44.491108 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerStarted","Data":"5952196ba6df1c3f6304cf54f5ef28bf948ecce04620fa732f4d7ee5016de208"} Dec 05 08:37:44 crc kubenswrapper[4876]: I1205 08:37:44.581103 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:37:44 crc kubenswrapper[4876]: I1205 08:37:44.581450 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerName="glance-httpd" containerID="cri-o://24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6" gracePeriod=30 Dec 05 08:37:44 crc kubenswrapper[4876]: I1205 08:37:44.581408 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerName="glance-log" containerID="cri-o://967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3" gracePeriod=30 Dec 05 08:37:45 crc kubenswrapper[4876]: I1205 08:37:45.526502 4876 generic.go:334] "Generic (PLEG): container finished" podID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerID="967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3" exitCode=143 Dec 05 08:37:45 crc kubenswrapper[4876]: I1205 08:37:45.526579 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c328d736-74fe-4075-b291-ea6ff38ef1ef","Type":"ContainerDied","Data":"967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3"} Dec 05 08:37:46 crc kubenswrapper[4876]: I1205 08:37:46.539598 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerStarted","Data":"1bdbdee706423edebbd985af870be9de4976047033cb18d8db36b228dd48d28e"} Dec 05 08:37:46 crc kubenswrapper[4876]: I1205 08:37:46.539937 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:37:46 crc kubenswrapper[4876]: I1205 08:37:46.571859 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.937205694 podStartE2EDuration="6.57183598s" podCreationTimestamp="2025-12-05 08:37:40 +0000 UTC" firstStartedPulling="2025-12-05 08:37:41.91567742 +0000 UTC m=+1326.404342042" lastFinishedPulling="2025-12-05 08:37:45.550307706 +0000 UTC m=+1330.038972328" observedRunningTime="2025-12-05 08:37:46.557581997 +0000 UTC m=+1331.046246619" watchObservedRunningTime="2025-12-05 08:37:46.57183598 +0000 UTC m=+1331.060500602" Dec 05 08:37:47 crc kubenswrapper[4876]: I1205 08:37:47.076527 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:37:47 crc kubenswrapper[4876]: I1205 08:37:47.547595 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerName="glance-log" containerID="cri-o://4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731" gracePeriod=30 Dec 05 08:37:47 crc kubenswrapper[4876]: I1205 08:37:47.547689 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerName="glance-httpd" containerID="cri-o://cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053" gracePeriod=30 Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.251547 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.342924 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.378192 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-combined-ca-bundle\") pod \"c328d736-74fe-4075-b291-ea6ff38ef1ef\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.378233 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-scripts\") pod \"c328d736-74fe-4075-b291-ea6ff38ef1ef\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.378272 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-logs\") pod \"c328d736-74fe-4075-b291-ea6ff38ef1ef\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.378334 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-httpd-run\") pod \"c328d736-74fe-4075-b291-ea6ff38ef1ef\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.378431 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xphnv\" (UniqueName: \"kubernetes.io/projected/c328d736-74fe-4075-b291-ea6ff38ef1ef-kube-api-access-xphnv\") pod \"c328d736-74fe-4075-b291-ea6ff38ef1ef\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.378454 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-public-tls-certs\") pod \"c328d736-74fe-4075-b291-ea6ff38ef1ef\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.378470 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"c328d736-74fe-4075-b291-ea6ff38ef1ef\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.378509 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-config-data\") pod \"c328d736-74fe-4075-b291-ea6ff38ef1ef\" (UID: \"c328d736-74fe-4075-b291-ea6ff38ef1ef\") " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.379465 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c328d736-74fe-4075-b291-ea6ff38ef1ef" (UID: "c328d736-74fe-4075-b291-ea6ff38ef1ef"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.379637 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-logs" (OuterVolumeSpecName: "logs") pod "c328d736-74fe-4075-b291-ea6ff38ef1ef" (UID: "c328d736-74fe-4075-b291-ea6ff38ef1ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.380295 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.380318 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c328d736-74fe-4075-b291-ea6ff38ef1ef-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.386310 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c328d736-74fe-4075-b291-ea6ff38ef1ef-kube-api-access-xphnv" (OuterVolumeSpecName: "kube-api-access-xphnv") pod "c328d736-74fe-4075-b291-ea6ff38ef1ef" (UID: "c328d736-74fe-4075-b291-ea6ff38ef1ef"). InnerVolumeSpecName "kube-api-access-xphnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.390226 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "c328d736-74fe-4075-b291-ea6ff38ef1ef" (UID: "c328d736-74fe-4075-b291-ea6ff38ef1ef"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.404144 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-scripts" (OuterVolumeSpecName: "scripts") pod "c328d736-74fe-4075-b291-ea6ff38ef1ef" (UID: "c328d736-74fe-4075-b291-ea6ff38ef1ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.406505 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c328d736-74fe-4075-b291-ea6ff38ef1ef" (UID: "c328d736-74fe-4075-b291-ea6ff38ef1ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.481600 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xphnv\" (UniqueName: \"kubernetes.io/projected/c328d736-74fe-4075-b291-ea6ff38ef1ef-kube-api-access-xphnv\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.481647 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.481657 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.481666 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.492220 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-config-data" (OuterVolumeSpecName: "config-data") pod "c328d736-74fe-4075-b291-ea6ff38ef1ef" (UID: "c328d736-74fe-4075-b291-ea6ff38ef1ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.498071 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c328d736-74fe-4075-b291-ea6ff38ef1ef" (UID: "c328d736-74fe-4075-b291-ea6ff38ef1ef"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.506593 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.560574 4876 generic.go:334] "Generic (PLEG): container finished" podID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerID="4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731" exitCode=143 Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.560679 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69166e3f-0a62-4a36-9fd4-2a78e35c43b6","Type":"ContainerDied","Data":"4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731"} Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.563148 4876 generic.go:334] "Generic (PLEG): container finished" podID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerID="24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6" exitCode=0 Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.563199 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c328d736-74fe-4075-b291-ea6ff38ef1ef","Type":"ContainerDied","Data":"24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6"} Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.563244 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.563275 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c328d736-74fe-4075-b291-ea6ff38ef1ef","Type":"ContainerDied","Data":"94f9a599f553ae6e0eab06a7f024790e0660e107db9042267aeb67e6f4fb61ac"} Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.563308 4876 scope.go:117] "RemoveContainer" containerID="24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.563869 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="sg-core" containerID="cri-o://6f33d42dce2006d056f4f4006958d70a629e5a33a5d4ef461913f97da2de3b03" gracePeriod=30 Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.563925 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="ceilometer-notification-agent" containerID="cri-o://5952196ba6df1c3f6304cf54f5ef28bf948ecce04620fa732f4d7ee5016de208" gracePeriod=30 Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.564030 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="ceilometer-central-agent" containerID="cri-o://2174f55d3c8fb34d9ec8b026c7acacf021fe0b3575d775c75c5bff54a352eff9" gracePeriod=30 Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.563937 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="proxy-httpd" containerID="cri-o://1bdbdee706423edebbd985af870be9de4976047033cb18d8db36b228dd48d28e" gracePeriod=30 Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.583027 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.583061 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.583073 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c328d736-74fe-4075-b291-ea6ff38ef1ef-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.602360 4876 scope.go:117] "RemoveContainer" containerID="967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.621384 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.631612 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.646857 4876 scope.go:117] "RemoveContainer" containerID="24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.647085 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:37:48 crc kubenswrapper[4876]: E1205 08:37:48.647515 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerName="glance-httpd" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.647537 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerName="glance-httpd" Dec 05 08:37:48 crc kubenswrapper[4876]: E1205 08:37:48.647588 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerName="glance-log" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.647599 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerName="glance-log" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.647816 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerName="glance-log" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.647847 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" containerName="glance-httpd" Dec 05 08:37:48 crc kubenswrapper[4876]: E1205 08:37:48.648402 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6\": container with ID starting with 24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6 not found: ID does not exist" containerID="24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.648451 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6"} err="failed to get container status \"24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6\": rpc error: code = NotFound desc = could not find container \"24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6\": container with ID starting with 24ed84016474a4f0a26aa017c9d9c045e6085321f8c36d48c2cb5f739e3905a6 not found: ID does not exist" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.648472 4876 scope.go:117] "RemoveContainer" containerID="967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3" Dec 05 08:37:48 crc kubenswrapper[4876]: E1205 08:37:48.648786 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3\": container with ID starting with 967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3 not found: ID does not exist" containerID="967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.648806 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3"} err="failed to get container status \"967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3\": rpc error: code = NotFound desc = could not find container \"967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3\": container with ID starting with 967c253c694f11dbbdf67424514d2fa99913f50d1018a6f17d15d0f8b4a444d3 not found: ID does not exist" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.650069 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.655333 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.655947 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.669683 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.787317 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.787388 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-scripts\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.787432 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.787467 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-logs\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.787582 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.787661 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-config-data\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.787705 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr6kx\" (UniqueName: \"kubernetes.io/projected/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-kube-api-access-dr6kx\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.787791 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.889391 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.889503 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.889558 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-scripts\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.889582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.889636 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-logs\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.889769 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.889846 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-config-data\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.889909 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr6kx\" (UniqueName: \"kubernetes.io/projected/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-kube-api-access-dr6kx\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.890018 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.890370 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-logs\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.890591 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.894432 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-scripts\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.897532 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.916159 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.927954 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.929831 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr6kx\" (UniqueName: \"kubernetes.io/projected/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-kube-api-access-dr6kx\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:48 crc kubenswrapper[4876]: I1205 08:37:48.931276 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e70a9dae-3695-42fa-bab0-b0ef8fc44e13-config-data\") pod \"glance-default-external-api-0\" (UID: \"e70a9dae-3695-42fa-bab0-b0ef8fc44e13\") " pod="openstack/glance-default-external-api-0" Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.097347 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.576459 4876 generic.go:334] "Generic (PLEG): container finished" podID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerID="1bdbdee706423edebbd985af870be9de4976047033cb18d8db36b228dd48d28e" exitCode=0 Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.576738 4876 generic.go:334] "Generic (PLEG): container finished" podID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerID="6f33d42dce2006d056f4f4006958d70a629e5a33a5d4ef461913f97da2de3b03" exitCode=2 Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.576746 4876 generic.go:334] "Generic (PLEG): container finished" podID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerID="5952196ba6df1c3f6304cf54f5ef28bf948ecce04620fa732f4d7ee5016de208" exitCode=0 Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.576763 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerDied","Data":"1bdbdee706423edebbd985af870be9de4976047033cb18d8db36b228dd48d28e"} Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.576785 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerDied","Data":"6f33d42dce2006d056f4f4006958d70a629e5a33a5d4ef461913f97da2de3b03"} Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.576795 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerDied","Data":"5952196ba6df1c3f6304cf54f5ef28bf948ecce04620fa732f4d7ee5016de208"} Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.696160 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 08:37:49 crc kubenswrapper[4876]: W1205 08:37:49.698567 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode70a9dae_3695_42fa_bab0_b0ef8fc44e13.slice/crio-e54e564c634eea2ff168335d9b7dccbe6cc88b2bcc0bebb5108af5984f7f8956 WatchSource:0}: Error finding container e54e564c634eea2ff168335d9b7dccbe6cc88b2bcc0bebb5108af5984f7f8956: Status 404 returned error can't find the container with id e54e564c634eea2ff168335d9b7dccbe6cc88b2bcc0bebb5108af5984f7f8956 Dec 05 08:37:49 crc kubenswrapper[4876]: I1205 08:37:49.861861 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c328d736-74fe-4075-b291-ea6ff38ef1ef" path="/var/lib/kubelet/pods/c328d736-74fe-4075-b291-ea6ff38ef1ef/volumes" Dec 05 08:37:50 crc kubenswrapper[4876]: I1205 08:37:50.599066 4876 generic.go:334] "Generic (PLEG): container finished" podID="99d7f1cf-79c4-42c4-bd75-a6588e8daaec" containerID="f87765f45920f1ea4498d1234b240f24872ab33ead370a56069f6fe647c0531e" exitCode=0 Dec 05 08:37:50 crc kubenswrapper[4876]: I1205 08:37:50.599180 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-54h52" event={"ID":"99d7f1cf-79c4-42c4-bd75-a6588e8daaec","Type":"ContainerDied","Data":"f87765f45920f1ea4498d1234b240f24872ab33ead370a56069f6fe647c0531e"} Dec 05 08:37:50 crc kubenswrapper[4876]: I1205 08:37:50.605357 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e70a9dae-3695-42fa-bab0-b0ef8fc44e13","Type":"ContainerStarted","Data":"1a07710c3b372dc4c37df524a84dc3bb04611409f4fcf2e79fd5a5feaf6696d1"} Dec 05 08:37:50 crc kubenswrapper[4876]: I1205 08:37:50.605409 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e70a9dae-3695-42fa-bab0-b0ef8fc44e13","Type":"ContainerStarted","Data":"e54e564c634eea2ff168335d9b7dccbe6cc88b2bcc0bebb5108af5984f7f8956"} Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.313751 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451356 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-logs\") pod \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451442 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-httpd-run\") pod \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451499 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-combined-ca-bundle\") pod \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451532 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-config-data\") pod \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451613 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-scripts\") pod \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451656 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-internal-tls-certs\") pod \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451679 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m97fp\" (UniqueName: \"kubernetes.io/projected/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-kube-api-access-m97fp\") pod \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451785 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\" (UID: \"69166e3f-0a62-4a36-9fd4-2a78e35c43b6\") " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451879 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "69166e3f-0a62-4a36-9fd4-2a78e35c43b6" (UID: "69166e3f-0a62-4a36-9fd4-2a78e35c43b6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.451956 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-logs" (OuterVolumeSpecName: "logs") pod "69166e3f-0a62-4a36-9fd4-2a78e35c43b6" (UID: "69166e3f-0a62-4a36-9fd4-2a78e35c43b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.452433 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.452456 4876 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.476649 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-kube-api-access-m97fp" (OuterVolumeSpecName: "kube-api-access-m97fp") pod "69166e3f-0a62-4a36-9fd4-2a78e35c43b6" (UID: "69166e3f-0a62-4a36-9fd4-2a78e35c43b6"). InnerVolumeSpecName "kube-api-access-m97fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.479579 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "69166e3f-0a62-4a36-9fd4-2a78e35c43b6" (UID: "69166e3f-0a62-4a36-9fd4-2a78e35c43b6"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.484467 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-scripts" (OuterVolumeSpecName: "scripts") pod "69166e3f-0a62-4a36-9fd4-2a78e35c43b6" (UID: "69166e3f-0a62-4a36-9fd4-2a78e35c43b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.490840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69166e3f-0a62-4a36-9fd4-2a78e35c43b6" (UID: "69166e3f-0a62-4a36-9fd4-2a78e35c43b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.516458 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "69166e3f-0a62-4a36-9fd4-2a78e35c43b6" (UID: "69166e3f-0a62-4a36-9fd4-2a78e35c43b6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.523885 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-config-data" (OuterVolumeSpecName: "config-data") pod "69166e3f-0a62-4a36-9fd4-2a78e35c43b6" (UID: "69166e3f-0a62-4a36-9fd4-2a78e35c43b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.554707 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.554759 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.554774 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.554786 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.554799 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m97fp\" (UniqueName: \"kubernetes.io/projected/69166e3f-0a62-4a36-9fd4-2a78e35c43b6-kube-api-access-m97fp\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.554844 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.586571 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.617572 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e70a9dae-3695-42fa-bab0-b0ef8fc44e13","Type":"ContainerStarted","Data":"1464a87d6dd567baf763b8918b20bd5fa3ff562ba0dcab78b065c639c3d536c6"} Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.620859 4876 generic.go:334] "Generic (PLEG): container finished" podID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerID="cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053" exitCode=0 Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.620942 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.620973 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69166e3f-0a62-4a36-9fd4-2a78e35c43b6","Type":"ContainerDied","Data":"cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053"} Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.621787 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69166e3f-0a62-4a36-9fd4-2a78e35c43b6","Type":"ContainerDied","Data":"3c1e3f37b7198dd537189638ecde00a7c086558583161c75e8dd05fba52f23fc"} Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.621810 4876 scope.go:117] "RemoveContainer" containerID="cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.651139 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.651113794 podStartE2EDuration="3.651113794s" podCreationTimestamp="2025-12-05 08:37:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:51.638768682 +0000 UTC m=+1336.127433304" watchObservedRunningTime="2025-12-05 08:37:51.651113794 +0000 UTC m=+1336.139778416" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.657749 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.677677 4876 scope.go:117] "RemoveContainer" containerID="4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.677884 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.693667 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.715654 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:37:51 crc kubenswrapper[4876]: E1205 08:37:51.716176 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerName="glance-httpd" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.716201 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerName="glance-httpd" Dec 05 08:37:51 crc kubenswrapper[4876]: E1205 08:37:51.716235 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerName="glance-log" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.716243 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerName="glance-log" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.716454 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerName="glance-log" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.716495 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" containerName="glance-httpd" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.718729 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.722188 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.722391 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.724776 4876 scope.go:117] "RemoveContainer" containerID="cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053" Dec 05 08:37:51 crc kubenswrapper[4876]: E1205 08:37:51.725222 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053\": container with ID starting with cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053 not found: ID does not exist" containerID="cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.725256 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053"} err="failed to get container status \"cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053\": rpc error: code = NotFound desc = could not find container \"cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053\": container with ID starting with cbb30285c1c6bdb2f1419b41f636ad24c2367667a5bb43486852ed7761dba053 not found: ID does not exist" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.725287 4876 scope.go:117] "RemoveContainer" containerID="4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731" Dec 05 08:37:51 crc kubenswrapper[4876]: E1205 08:37:51.726073 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731\": container with ID starting with 4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731 not found: ID does not exist" containerID="4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.726119 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731"} err="failed to get container status \"4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731\": rpc error: code = NotFound desc = could not find container \"4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731\": container with ID starting with 4e438f24a73e70c4ed939e547fe4dc89013ec4b70050e3d2d19c3aaa055f7731 not found: ID does not exist" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.730294 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.837505 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69166e3f-0a62-4a36-9fd4-2a78e35c43b6" path="/var/lib/kubelet/pods/69166e3f-0a62-4a36-9fd4-2a78e35c43b6/volumes" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.862748 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00302ae2-a0ed-4a18-80e4-67ac2fed0230-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.862920 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.862975 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.863048 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.863092 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-config-data\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.863116 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-scripts\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.863192 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00302ae2-a0ed-4a18-80e4-67ac2fed0230-logs\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.863315 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66ntm\" (UniqueName: \"kubernetes.io/projected/00302ae2-a0ed-4a18-80e4-67ac2fed0230-kube-api-access-66ntm\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.968864 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.968924 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-config-data\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.968956 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-scripts\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.968997 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00302ae2-a0ed-4a18-80e4-67ac2fed0230-logs\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.969055 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66ntm\" (UniqueName: \"kubernetes.io/projected/00302ae2-a0ed-4a18-80e4-67ac2fed0230-kube-api-access-66ntm\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.969093 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00302ae2-a0ed-4a18-80e4-67ac2fed0230-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.969142 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.969174 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.969558 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.973018 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00302ae2-a0ed-4a18-80e4-67ac2fed0230-logs\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.973197 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/00302ae2-a0ed-4a18-80e4-67ac2fed0230-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.976629 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-scripts\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.985281 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-config-data\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:51 crc kubenswrapper[4876]: I1205 08:37:51.989665 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.007992 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00302ae2-a0ed-4a18-80e4-67ac2fed0230-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.008423 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66ntm\" (UniqueName: \"kubernetes.io/projected/00302ae2-a0ed-4a18-80e4-67ac2fed0230-kube-api-access-66ntm\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.038237 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"00302ae2-a0ed-4a18-80e4-67ac2fed0230\") " pod="openstack/glance-default-internal-api-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.176969 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.273138 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbkxt\" (UniqueName: \"kubernetes.io/projected/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-kube-api-access-wbkxt\") pod \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.273211 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-config-data\") pod \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.273289 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-scripts\") pod \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.273400 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-combined-ca-bundle\") pod \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\" (UID: \"99d7f1cf-79c4-42c4-bd75-a6588e8daaec\") " Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.280097 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-kube-api-access-wbkxt" (OuterVolumeSpecName: "kube-api-access-wbkxt") pod "99d7f1cf-79c4-42c4-bd75-a6588e8daaec" (UID: "99d7f1cf-79c4-42c4-bd75-a6588e8daaec"). InnerVolumeSpecName "kube-api-access-wbkxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.282961 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-scripts" (OuterVolumeSpecName: "scripts") pod "99d7f1cf-79c4-42c4-bd75-a6588e8daaec" (UID: "99d7f1cf-79c4-42c4-bd75-a6588e8daaec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.298860 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99d7f1cf-79c4-42c4-bd75-a6588e8daaec" (UID: "99d7f1cf-79c4-42c4-bd75-a6588e8daaec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.315481 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-config-data" (OuterVolumeSpecName: "config-data") pod "99d7f1cf-79c4-42c4-bd75-a6588e8daaec" (UID: "99d7f1cf-79c4-42c4-bd75-a6588e8daaec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.336761 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.375419 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.375454 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbkxt\" (UniqueName: \"kubernetes.io/projected/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-kube-api-access-wbkxt\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.375466 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.375476 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d7f1cf-79c4-42c4-bd75-a6588e8daaec-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.635099 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-54h52" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.635098 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-54h52" event={"ID":"99d7f1cf-79c4-42c4-bd75-a6588e8daaec","Type":"ContainerDied","Data":"74fe2a9c92335a45e2653074cc18963f99ac6d873205236a31e858f1e3a2f583"} Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.635553 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74fe2a9c92335a45e2653074cc18963f99ac6d873205236a31e858f1e3a2f583" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.768025 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:37:52 crc kubenswrapper[4876]: E1205 08:37:52.768370 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d7f1cf-79c4-42c4-bd75-a6588e8daaec" containerName="nova-cell0-conductor-db-sync" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.768387 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d7f1cf-79c4-42c4-bd75-a6588e8daaec" containerName="nova-cell0-conductor-db-sync" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.768577 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d7f1cf-79c4-42c4-bd75-a6588e8daaec" containerName="nova-cell0-conductor-db-sync" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.769154 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.771818 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-x7tb8" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.771983 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.781382 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.888663 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppzbt\" (UniqueName: \"kubernetes.io/projected/1223251d-7186-4aa3-8d6a-f19b3f248512-kube-api-access-ppzbt\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.888803 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.888831 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.902053 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 08:37:52 crc kubenswrapper[4876]: W1205 08:37:52.904528 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00302ae2_a0ed_4a18_80e4_67ac2fed0230.slice/crio-d4cbd72c74c5e0d9855a7bb644846c4324f5643f88f47f54c7ce7ae80b0b4b32 WatchSource:0}: Error finding container d4cbd72c74c5e0d9855a7bb644846c4324f5643f88f47f54c7ce7ae80b0b4b32: Status 404 returned error can't find the container with id d4cbd72c74c5e0d9855a7bb644846c4324f5643f88f47f54c7ce7ae80b0b4b32 Dec 05 08:37:52 crc kubenswrapper[4876]: I1205 08:37:52.990869 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppzbt\" (UniqueName: \"kubernetes.io/projected/1223251d-7186-4aa3-8d6a-f19b3f248512-kube-api-access-ppzbt\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:52.991010 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.081744 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.087051 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.093683 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.125870 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppzbt\" (UniqueName: \"kubernetes.io/projected/1223251d-7186-4aa3-8d6a-f19b3f248512-kube-api-access-ppzbt\") pod \"nova-cell0-conductor-0\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.395609 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.687433 4876 generic.go:334] "Generic (PLEG): container finished" podID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerID="2174f55d3c8fb34d9ec8b026c7acacf021fe0b3575d775c75c5bff54a352eff9" exitCode=0 Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.687807 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerDied","Data":"2174f55d3c8fb34d9ec8b026c7acacf021fe0b3575d775c75c5bff54a352eff9"} Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.707289 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"00302ae2-a0ed-4a18-80e4-67ac2fed0230","Type":"ContainerStarted","Data":"2d10e6dfef0f86320d2d4b83031bd75b6d000de991e1f3cc16a66395c9287837"} Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.707361 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"00302ae2-a0ed-4a18-80e4-67ac2fed0230","Type":"ContainerStarted","Data":"d4cbd72c74c5e0d9855a7bb644846c4324f5643f88f47f54c7ce7ae80b0b4b32"} Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.822828 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:53 crc kubenswrapper[4876]: I1205 08:37:53.994817 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.002717 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-scripts\") pod \"aa99f31a-f42e-43c0-b104-b484332aa37e\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.002915 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-config-data\") pod \"aa99f31a-f42e-43c0-b104-b484332aa37e\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.002998 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-run-httpd\") pod \"aa99f31a-f42e-43c0-b104-b484332aa37e\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.003076 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-combined-ca-bundle\") pod \"aa99f31a-f42e-43c0-b104-b484332aa37e\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.003120 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9djcj\" (UniqueName: \"kubernetes.io/projected/aa99f31a-f42e-43c0-b104-b484332aa37e-kube-api-access-9djcj\") pod \"aa99f31a-f42e-43c0-b104-b484332aa37e\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.003168 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-sg-core-conf-yaml\") pod \"aa99f31a-f42e-43c0-b104-b484332aa37e\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.003236 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-log-httpd\") pod \"aa99f31a-f42e-43c0-b104-b484332aa37e\" (UID: \"aa99f31a-f42e-43c0-b104-b484332aa37e\") " Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.005029 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aa99f31a-f42e-43c0-b104-b484332aa37e" (UID: "aa99f31a-f42e-43c0-b104-b484332aa37e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.011214 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aa99f31a-f42e-43c0-b104-b484332aa37e" (UID: "aa99f31a-f42e-43c0-b104-b484332aa37e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.012027 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-scripts" (OuterVolumeSpecName: "scripts") pod "aa99f31a-f42e-43c0-b104-b484332aa37e" (UID: "aa99f31a-f42e-43c0-b104-b484332aa37e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:54 crc kubenswrapper[4876]: W1205 08:37:54.014397 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1223251d_7186_4aa3_8d6a_f19b3f248512.slice/crio-1d16eeb0818d6f0ee1afcaf230ffa17b340df18d3d188fa88740aa6e4058a363 WatchSource:0}: Error finding container 1d16eeb0818d6f0ee1afcaf230ffa17b340df18d3d188fa88740aa6e4058a363: Status 404 returned error can't find the container with id 1d16eeb0818d6f0ee1afcaf230ffa17b340df18d3d188fa88740aa6e4058a363 Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.014970 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa99f31a-f42e-43c0-b104-b484332aa37e-kube-api-access-9djcj" (OuterVolumeSpecName: "kube-api-access-9djcj") pod "aa99f31a-f42e-43c0-b104-b484332aa37e" (UID: "aa99f31a-f42e-43c0-b104-b484332aa37e"). InnerVolumeSpecName "kube-api-access-9djcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.038634 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aa99f31a-f42e-43c0-b104-b484332aa37e" (UID: "aa99f31a-f42e-43c0-b104-b484332aa37e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.105298 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9djcj\" (UniqueName: \"kubernetes.io/projected/aa99f31a-f42e-43c0-b104-b484332aa37e-kube-api-access-9djcj\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.105332 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.105341 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.105350 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.105358 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa99f31a-f42e-43c0-b104-b484332aa37e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.133205 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa99f31a-f42e-43c0-b104-b484332aa37e" (UID: "aa99f31a-f42e-43c0-b104-b484332aa37e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.142016 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-config-data" (OuterVolumeSpecName: "config-data") pod "aa99f31a-f42e-43c0-b104-b484332aa37e" (UID: "aa99f31a-f42e-43c0-b104-b484332aa37e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.207545 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.207574 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa99f31a-f42e-43c0-b104-b484332aa37e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.720986 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.721028 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa99f31a-f42e-43c0-b104-b484332aa37e","Type":"ContainerDied","Data":"9f2eff702d021d79aea80826e0dc7eb8b92562790db99a759f03c86e1e1d76c3"} Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.722216 4876 scope.go:117] "RemoveContainer" containerID="1bdbdee706423edebbd985af870be9de4976047033cb18d8db36b228dd48d28e" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.723704 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"00302ae2-a0ed-4a18-80e4-67ac2fed0230","Type":"ContainerStarted","Data":"d0396e9359277cf33a08c3107e6c2ea5a5abe0b51456f72efb47816845227042"} Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.730504 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1223251d-7186-4aa3-8d6a-f19b3f248512","Type":"ContainerStarted","Data":"6f598c51ab6b00be740b404712f1b796dd2890b2479dcbfb419971c8a29c2989"} Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.730545 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1223251d-7186-4aa3-8d6a-f19b3f248512","Type":"ContainerStarted","Data":"1d16eeb0818d6f0ee1afcaf230ffa17b340df18d3d188fa88740aa6e4058a363"} Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.730828 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.755784 4876 scope.go:117] "RemoveContainer" containerID="6f33d42dce2006d056f4f4006958d70a629e5a33a5d4ef461913f97da2de3b03" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.760707 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.760684754 podStartE2EDuration="3.760684754s" podCreationTimestamp="2025-12-05 08:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:54.756139252 +0000 UTC m=+1339.244803874" watchObservedRunningTime="2025-12-05 08:37:54.760684754 +0000 UTC m=+1339.249349386" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.792426 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.784535345 podStartE2EDuration="2.784535345s" podCreationTimestamp="2025-12-05 08:37:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:37:54.776990232 +0000 UTC m=+1339.265654854" watchObservedRunningTime="2025-12-05 08:37:54.784535345 +0000 UTC m=+1339.273199967" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.811462 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.813628 4876 scope.go:117] "RemoveContainer" containerID="5952196ba6df1c3f6304cf54f5ef28bf948ecce04620fa732f4d7ee5016de208" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.824056 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832109 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:54 crc kubenswrapper[4876]: E1205 08:37:54.832533 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="ceilometer-notification-agent" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832555 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="ceilometer-notification-agent" Dec 05 08:37:54 crc kubenswrapper[4876]: E1205 08:37:54.832580 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="proxy-httpd" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832589 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="proxy-httpd" Dec 05 08:37:54 crc kubenswrapper[4876]: E1205 08:37:54.832616 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="ceilometer-central-agent" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832625 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="ceilometer-central-agent" Dec 05 08:37:54 crc kubenswrapper[4876]: E1205 08:37:54.832647 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="sg-core" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832655 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="sg-core" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832885 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="sg-core" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832933 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="ceilometer-central-agent" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832947 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="proxy-httpd" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.832964 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" containerName="ceilometer-notification-agent" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.835297 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.838753 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.839035 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.841512 4876 scope.go:117] "RemoveContainer" containerID="2174f55d3c8fb34d9ec8b026c7acacf021fe0b3575d775c75c5bff54a352eff9" Dec 05 08:37:54 crc kubenswrapper[4876]: I1205 08:37:54.842888 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.025041 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-log-httpd\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.025449 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.025628 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-scripts\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.025650 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.025682 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-config-data\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.025713 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-run-httpd\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.025894 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz6xg\" (UniqueName: \"kubernetes.io/projected/4e42626e-224c-4d7d-9fb9-5620eec12bd6-kube-api-access-gz6xg\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.127912 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz6xg\" (UniqueName: \"kubernetes.io/projected/4e42626e-224c-4d7d-9fb9-5620eec12bd6-kube-api-access-gz6xg\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.127994 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-log-httpd\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.128045 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.128161 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-scripts\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.128178 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.128201 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-config-data\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.128239 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-run-httpd\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.128972 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-run-httpd\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.132423 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-log-httpd\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.135292 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-scripts\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.142320 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-config-data\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.145776 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.158673 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.217805 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz6xg\" (UniqueName: \"kubernetes.io/projected/4e42626e-224c-4d7d-9fb9-5620eec12bd6-kube-api-access-gz6xg\") pod \"ceilometer-0\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.455710 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.841431 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa99f31a-f42e-43c0-b104-b484332aa37e" path="/var/lib/kubelet/pods/aa99f31a-f42e-43c0-b104-b484332aa37e/volumes" Dec 05 08:37:55 crc kubenswrapper[4876]: I1205 08:37:55.917613 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:55 crc kubenswrapper[4876]: W1205 08:37:55.926073 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e42626e_224c_4d7d_9fb9_5620eec12bd6.slice/crio-5d85d0720cd42c57c9253af57939bcc232afc39ec49569167828341dd726a851 WatchSource:0}: Error finding container 5d85d0720cd42c57c9253af57939bcc232afc39ec49569167828341dd726a851: Status 404 returned error can't find the container with id 5d85d0720cd42c57c9253af57939bcc232afc39ec49569167828341dd726a851 Dec 05 08:37:56 crc kubenswrapper[4876]: I1205 08:37:56.782025 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerStarted","Data":"ee35c63bf0ef3381f4062b3f5ad6422b377d497b01ad50879ee40ae96a91249d"} Dec 05 08:37:56 crc kubenswrapper[4876]: I1205 08:37:56.782502 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerStarted","Data":"5d85d0720cd42c57c9253af57939bcc232afc39ec49569167828341dd726a851"} Dec 05 08:37:57 crc kubenswrapper[4876]: I1205 08:37:57.350569 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:37:57 crc kubenswrapper[4876]: I1205 08:37:57.793841 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerStarted","Data":"10de5d8a5060a77202e9a1fe3db2ba99b2de638f1ec35936aaf78afc260bac80"} Dec 05 08:37:58 crc kubenswrapper[4876]: I1205 08:37:58.803936 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerStarted","Data":"36eb9f4032f915f789c6b1d147913e6ac6810d9c22e5c453d71c8eb4f9009f38"} Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.098115 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.098502 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.141830 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.156580 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.820811 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerStarted","Data":"27eb81dfb5009f33750db93707703ac69a87648a540afa8ad9bfa3a0bc98c445"} Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.821137 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.821073 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="ceilometer-central-agent" containerID="cri-o://ee35c63bf0ef3381f4062b3f5ad6422b377d497b01ad50879ee40ae96a91249d" gracePeriod=30 Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.821251 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="ceilometer-notification-agent" containerID="cri-o://10de5d8a5060a77202e9a1fe3db2ba99b2de638f1ec35936aaf78afc260bac80" gracePeriod=30 Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.821305 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="sg-core" containerID="cri-o://36eb9f4032f915f789c6b1d147913e6ac6810d9c22e5c453d71c8eb4f9009f38" gracePeriod=30 Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.821394 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.822078 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="proxy-httpd" containerID="cri-o://27eb81dfb5009f33750db93707703ac69a87648a540afa8ad9bfa3a0bc98c445" gracePeriod=30 Dec 05 08:37:59 crc kubenswrapper[4876]: I1205 08:37:59.850054 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.860790825 podStartE2EDuration="5.85003666s" podCreationTimestamp="2025-12-05 08:37:54 +0000 UTC" firstStartedPulling="2025-12-05 08:37:55.927714659 +0000 UTC m=+1340.416379281" lastFinishedPulling="2025-12-05 08:37:58.916960494 +0000 UTC m=+1343.405625116" observedRunningTime="2025-12-05 08:37:59.848971951 +0000 UTC m=+1344.337636583" watchObservedRunningTime="2025-12-05 08:37:59.85003666 +0000 UTC m=+1344.338701282" Dec 05 08:38:00 crc kubenswrapper[4876]: I1205 08:38:00.847803 4876 generic.go:334] "Generic (PLEG): container finished" podID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerID="27eb81dfb5009f33750db93707703ac69a87648a540afa8ad9bfa3a0bc98c445" exitCode=0 Dec 05 08:38:00 crc kubenswrapper[4876]: I1205 08:38:00.848167 4876 generic.go:334] "Generic (PLEG): container finished" podID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerID="36eb9f4032f915f789c6b1d147913e6ac6810d9c22e5c453d71c8eb4f9009f38" exitCode=2 Dec 05 08:38:00 crc kubenswrapper[4876]: I1205 08:38:00.848182 4876 generic.go:334] "Generic (PLEG): container finished" podID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerID="10de5d8a5060a77202e9a1fe3db2ba99b2de638f1ec35936aaf78afc260bac80" exitCode=0 Dec 05 08:38:00 crc kubenswrapper[4876]: I1205 08:38:00.847882 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerDied","Data":"27eb81dfb5009f33750db93707703ac69a87648a540afa8ad9bfa3a0bc98c445"} Dec 05 08:38:00 crc kubenswrapper[4876]: I1205 08:38:00.849234 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerDied","Data":"36eb9f4032f915f789c6b1d147913e6ac6810d9c22e5c453d71c8eb4f9009f38"} Dec 05 08:38:00 crc kubenswrapper[4876]: I1205 08:38:00.849253 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerDied","Data":"10de5d8a5060a77202e9a1fe3db2ba99b2de638f1ec35936aaf78afc260bac80"} Dec 05 08:38:01 crc kubenswrapper[4876]: I1205 08:38:01.862762 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 08:38:01 crc kubenswrapper[4876]: I1205 08:38:01.863329 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 08:38:01 crc kubenswrapper[4876]: I1205 08:38:01.894616 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 08:38:02 crc kubenswrapper[4876]: I1205 08:38:02.337356 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 08:38:02 crc kubenswrapper[4876]: I1205 08:38:02.337700 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 08:38:02 crc kubenswrapper[4876]: I1205 08:38:02.375326 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 08:38:02 crc kubenswrapper[4876]: I1205 08:38:02.389740 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 08:38:02 crc kubenswrapper[4876]: I1205 08:38:02.869567 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 08:38:02 crc kubenswrapper[4876]: I1205 08:38:02.869637 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 08:38:03 crc kubenswrapper[4876]: I1205 08:38:03.426745 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:03 crc kubenswrapper[4876]: I1205 08:38:03.909892 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:38:03 crc kubenswrapper[4876]: I1205 08:38:03.910400 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="1223251d-7186-4aa3-8d6a-f19b3f248512" containerName="nova-cell0-conductor-conductor" containerID="cri-o://6f598c51ab6b00be740b404712f1b796dd2890b2479dcbfb419971c8a29c2989" gracePeriod=30 Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.056917 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vhxdf"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.058179 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.060362 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.063424 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.068137 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vhxdf"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.135346 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-scripts\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.135393 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bnfb\" (UniqueName: \"kubernetes.io/projected/b6629f58-7171-4cd4-a018-6b9c40a8bfee-kube-api-access-7bnfb\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.135543 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-config-data\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.135583 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.251681 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-scripts\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.252141 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bnfb\" (UniqueName: \"kubernetes.io/projected/b6629f58-7171-4cd4-a018-6b9c40a8bfee-kube-api-access-7bnfb\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.252218 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-config-data\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.252240 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.269281 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-config-data\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.276918 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.277435 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-scripts\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.309483 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bnfb\" (UniqueName: \"kubernetes.io/projected/b6629f58-7171-4cd4-a018-6b9c40a8bfee-kube-api-access-7bnfb\") pod \"nova-cell0-cell-mapping-vhxdf\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.342067 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.344104 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.358865 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.380570 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.406848 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.467461 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.467546 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-config-data\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.467584 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdlz4\" (UniqueName: \"kubernetes.io/projected/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-kube-api-access-bdlz4\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.467659 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-logs\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.467749 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.469172 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.484378 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.562443 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.574969 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.575053 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-config-data\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.575086 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdlz4\" (UniqueName: \"kubernetes.io/projected/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-kube-api-access-bdlz4\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.575161 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6nbq\" (UniqueName: \"kubernetes.io/projected/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-kube-api-access-x6nbq\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.575185 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-config-data\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.575210 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.575232 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-logs\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.575816 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-logs\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.579797 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.594158 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-config-data\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.609391 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.611432 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.617169 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdlz4\" (UniqueName: \"kubernetes.io/projected/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-kube-api-access-bdlz4\") pod \"nova-api-0\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.618243 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.619584 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.650960 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.662814 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.664025 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.666804 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.667834 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.677712 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6nbq\" (UniqueName: \"kubernetes.io/projected/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-kube-api-access-x6nbq\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.677768 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-config-data\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.677807 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.683944 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-config-data\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.689335 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.716612 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6nbq\" (UniqueName: \"kubernetes.io/projected/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-kube-api-access-x6nbq\") pod \"nova-scheduler-0\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.785827 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7877d89589-pw6bc"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.787188 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.788720 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-config-data\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.788864 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntg69\" (UniqueName: \"kubernetes.io/projected/9291e5f9-2b2d-4785-bdb4-0e43c997f458-kube-api-access-ntg69\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.788913 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64660a67-5194-45d4-a364-228db58cf2a9-logs\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.788931 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.788947 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr5sb\" (UniqueName: \"kubernetes.io/projected/64660a67-5194-45d4-a364-228db58cf2a9-kube-api-access-mr5sb\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.788984 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.789055 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.870377 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-pw6bc"] Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.926805 4876 generic.go:334] "Generic (PLEG): container finished" podID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerID="ee35c63bf0ef3381f4062b3f5ad6422b377d497b01ad50879ee40ae96a91249d" exitCode=0 Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.927182 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerDied","Data":"ee35c63bf0ef3381f4062b3f5ad6422b377d497b01ad50879ee40ae96a91249d"} Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.978519 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.979074 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.979230 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-svc\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.979293 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.979392 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.979441 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.979492 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxzmf\" (UniqueName: \"kubernetes.io/projected/14bc9681-6da3-4e8b-8a88-4940f7b359de-kube-api-access-rxzmf\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.979604 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-config\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.979716 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-config-data\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.986211 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.986272 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntg69\" (UniqueName: \"kubernetes.io/projected/9291e5f9-2b2d-4785-bdb4-0e43c997f458-kube-api-access-ntg69\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.986389 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64660a67-5194-45d4-a364-228db58cf2a9-logs\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.986439 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.986467 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr5sb\" (UniqueName: \"kubernetes.io/projected/64660a67-5194-45d4-a364-228db58cf2a9-kube-api-access-mr5sb\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.989859 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:04 crc kubenswrapper[4876]: I1205 08:38:04.990343 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64660a67-5194-45d4-a364-228db58cf2a9-logs\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.028947 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.031646 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-config-data\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.034842 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.038120 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntg69\" (UniqueName: \"kubernetes.io/projected/9291e5f9-2b2d-4785-bdb4-0e43c997f458-kube-api-access-ntg69\") pod \"nova-cell1-novncproxy-0\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.048178 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr5sb\" (UniqueName: \"kubernetes.io/projected/64660a67-5194-45d4-a364-228db58cf2a9-kube-api-access-mr5sb\") pod \"nova-metadata-0\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " pod="openstack/nova-metadata-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.090054 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.090177 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-svc\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.090208 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.090240 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.090274 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxzmf\" (UniqueName: \"kubernetes.io/projected/14bc9681-6da3-4e8b-8a88-4940f7b359de-kube-api-access-rxzmf\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.090298 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-config\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.091242 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-config\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.091770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.092467 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-svc\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.093199 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.093770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.113603 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.119599 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxzmf\" (UniqueName: \"kubernetes.io/projected/14bc9681-6da3-4e8b-8a88-4940f7b359de-kube-api-access-rxzmf\") pod \"dnsmasq-dns-7877d89589-pw6bc\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.156042 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.190982 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-sg-core-conf-yaml\") pod \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.191037 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-log-httpd\") pod \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.191130 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-run-httpd\") pod \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.191145 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-scripts\") pod \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.191442 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4e42626e-224c-4d7d-9fb9-5620eec12bd6" (UID: "4e42626e-224c-4d7d-9fb9-5620eec12bd6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.191652 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4e42626e-224c-4d7d-9fb9-5620eec12bd6" (UID: "4e42626e-224c-4d7d-9fb9-5620eec12bd6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.192037 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-config-data\") pod \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.192070 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-combined-ca-bundle\") pod \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.192099 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz6xg\" (UniqueName: \"kubernetes.io/projected/4e42626e-224c-4d7d-9fb9-5620eec12bd6-kube-api-access-gz6xg\") pod \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\" (UID: \"4e42626e-224c-4d7d-9fb9-5620eec12bd6\") " Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.192477 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.192494 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4e42626e-224c-4d7d-9fb9-5620eec12bd6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.197991 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-scripts" (OuterVolumeSpecName: "scripts") pod "4e42626e-224c-4d7d-9fb9-5620eec12bd6" (UID: "4e42626e-224c-4d7d-9fb9-5620eec12bd6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.198102 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e42626e-224c-4d7d-9fb9-5620eec12bd6-kube-api-access-gz6xg" (OuterVolumeSpecName: "kube-api-access-gz6xg") pod "4e42626e-224c-4d7d-9fb9-5620eec12bd6" (UID: "4e42626e-224c-4d7d-9fb9-5620eec12bd6"). InnerVolumeSpecName "kube-api-access-gz6xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.239201 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4e42626e-224c-4d7d-9fb9-5620eec12bd6" (UID: "4e42626e-224c-4d7d-9fb9-5620eec12bd6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.295592 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.295630 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz6xg\" (UniqueName: \"kubernetes.io/projected/4e42626e-224c-4d7d-9fb9-5620eec12bd6-kube-api-access-gz6xg\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.295645 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.297650 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.334623 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.346117 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e42626e-224c-4d7d-9fb9-5620eec12bd6" (UID: "4e42626e-224c-4d7d-9fb9-5620eec12bd6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.397180 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.407992 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vhxdf"] Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.423050 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.460788 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-config-data" (OuterVolumeSpecName: "config-data") pod "4e42626e-224c-4d7d-9fb9-5620eec12bd6" (UID: "4e42626e-224c-4d7d-9fb9-5620eec12bd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.498559 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e42626e-224c-4d7d-9fb9-5620eec12bd6-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.620460 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.778833 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.778960 4876 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.847594 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-pw6bc"] Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.888012 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 08:38:05 crc kubenswrapper[4876]: I1205 08:38:05.967612 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f","Type":"ContainerStarted","Data":"ab1a3d93d2c56e19f2f087cfac518c932abc66011ebf6276ee48592b0e0a44f8"} Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.010226 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4e42626e-224c-4d7d-9fb9-5620eec12bd6","Type":"ContainerDied","Data":"5d85d0720cd42c57c9253af57939bcc232afc39ec49569167828341dd726a851"} Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.010282 4876 scope.go:117] "RemoveContainer" containerID="27eb81dfb5009f33750db93707703ac69a87648a540afa8ad9bfa3a0bc98c445" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.010420 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.020094 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.038964 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gpm57"] Dec 05 08:38:06 crc kubenswrapper[4876]: E1205 08:38:06.039469 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="sg-core" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.039512 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="sg-core" Dec 05 08:38:06 crc kubenswrapper[4876]: E1205 08:38:06.039534 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="ceilometer-notification-agent" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.039541 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="ceilometer-notification-agent" Dec 05 08:38:06 crc kubenswrapper[4876]: E1205 08:38:06.039597 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="ceilometer-central-agent" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.039606 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="ceilometer-central-agent" Dec 05 08:38:06 crc kubenswrapper[4876]: E1205 08:38:06.039638 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="proxy-httpd" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.039645 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="proxy-httpd" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.040039 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="ceilometer-notification-agent" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.040100 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="ceilometer-central-agent" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.040120 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="proxy-httpd" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.040137 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" containerName="sg-core" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.040941 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6eb8b37-115c-4d30-9358-c4773f3e1d9e","Type":"ContainerStarted","Data":"15fd688450ef0b4adaf229d73d34cda3f34f51191a4be635ce608e71c084f7fa"} Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.041088 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.044800 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.045002 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.052147 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gpm57"] Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.052618 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" event={"ID":"14bc9681-6da3-4e8b-8a88-4940f7b359de","Type":"ContainerStarted","Data":"5fbc934ad18e3a94648eb14a5f02e13db8c65f13ca458c53731e3665fd5f1ba1"} Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.056338 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vhxdf" event={"ID":"b6629f58-7171-4cd4-a018-6b9c40a8bfee","Type":"ContainerStarted","Data":"61ca35015ec9ca00cca71b6d6bd2a6cd1722b775213b9192131b407f5798c455"} Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.056365 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vhxdf" event={"ID":"b6629f58-7171-4cd4-a018-6b9c40a8bfee","Type":"ContainerStarted","Data":"068fb74cc90c5981f8cbeffab4a1dfdc90ad7c82a883eddb99c2651cc2bb018d"} Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.066217 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.084960 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.117000 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.119876 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.120163 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vhxdf" podStartSLOduration=2.120151959 podStartE2EDuration="2.120151959s" podCreationTimestamp="2025-12-05 08:38:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:06.097153971 +0000 UTC m=+1350.585818593" watchObservedRunningTime="2025-12-05 08:38:06.120151959 +0000 UTC m=+1350.608816581" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.124924 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.125099 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.132803 4876 scope.go:117] "RemoveContainer" containerID="36eb9f4032f915f789c6b1d147913e6ac6810d9c22e5c453d71c8eb4f9009f38" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.139588 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.182400 4876 scope.go:117] "RemoveContainer" containerID="10de5d8a5060a77202e9a1fe3db2ba99b2de638f1ec35936aaf78afc260bac80" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.197514 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:06 crc kubenswrapper[4876]: W1205 08:38:06.227938 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9291e5f9_2b2d_4785_bdb4_0e43c997f458.slice/crio-da55e836d4b01c08f93378809f388aba24cee1751e65419913f0b26b01fd3a6f WatchSource:0}: Error finding container da55e836d4b01c08f93378809f388aba24cee1751e65419913f0b26b01fd3a6f: Status 404 returned error can't find the container with id da55e836d4b01c08f93378809f388aba24cee1751e65419913f0b26b01fd3a6f Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.233791 4876 scope.go:117] "RemoveContainer" containerID="ee35c63bf0ef3381f4062b3f5ad6422b377d497b01ad50879ee40ae96a91249d" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.236259 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-scripts\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.236372 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqtkn\" (UniqueName: \"kubernetes.io/projected/4347cbb9-a742-4985-b4c8-a50c72d88a14-kube-api-access-gqtkn\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.236458 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.236742 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.236853 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-scripts\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.237416 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skz2r\" (UniqueName: \"kubernetes.io/projected/198b165f-0f97-4551-a939-73d07f958d01-kube-api-access-skz2r\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.237502 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-config-data\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.237572 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-log-httpd\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.237649 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-run-httpd\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.237737 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.237804 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-config-data\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.339853 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skz2r\" (UniqueName: \"kubernetes.io/projected/198b165f-0f97-4551-a939-73d07f958d01-kube-api-access-skz2r\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340246 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-config-data\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340275 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-log-httpd\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340302 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-run-httpd\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340354 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340386 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-config-data\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340411 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-scripts\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340443 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqtkn\" (UniqueName: \"kubernetes.io/projected/4347cbb9-a742-4985-b4c8-a50c72d88a14-kube-api-access-gqtkn\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340473 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340504 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.340578 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-scripts\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.344375 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-scripts\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.345868 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-log-httpd\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.346306 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.347241 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-run-httpd\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.349522 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-scripts\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.356612 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.356673 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.356859 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-config-data\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.356863 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-config-data\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.359546 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqtkn\" (UniqueName: \"kubernetes.io/projected/4347cbb9-a742-4985-b4c8-a50c72d88a14-kube-api-access-gqtkn\") pod \"ceilometer-0\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.364646 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skz2r\" (UniqueName: \"kubernetes.io/projected/198b165f-0f97-4551-a939-73d07f958d01-kube-api-access-skz2r\") pod \"nova-cell1-conductor-db-sync-gpm57\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.371522 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.464392 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:38:06 crc kubenswrapper[4876]: I1205 08:38:06.930524 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gpm57"] Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.083127 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gpm57" event={"ID":"198b165f-0f97-4551-a939-73d07f958d01","Type":"ContainerStarted","Data":"b3980706c4b98c83553537121c7a43e78596f47ff2222e49cae387ad8bb3af62"} Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.084251 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9291e5f9-2b2d-4785-bdb4-0e43c997f458","Type":"ContainerStarted","Data":"da55e836d4b01c08f93378809f388aba24cee1751e65419913f0b26b01fd3a6f"} Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.093181 4876 generic.go:334] "Generic (PLEG): container finished" podID="14bc9681-6da3-4e8b-8a88-4940f7b359de" containerID="6d5ddaea1f3d3216e45035a5b36287b4b0b2809faf1e49b87cd4d9df114e622d" exitCode=0 Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.093689 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" event={"ID":"14bc9681-6da3-4e8b-8a88-4940f7b359de","Type":"ContainerDied","Data":"6d5ddaea1f3d3216e45035a5b36287b4b0b2809faf1e49b87cd4d9df114e622d"} Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.101997 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.104835 4876 generic.go:334] "Generic (PLEG): container finished" podID="1223251d-7186-4aa3-8d6a-f19b3f248512" containerID="6f598c51ab6b00be740b404712f1b796dd2890b2479dcbfb419971c8a29c2989" exitCode=0 Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.104924 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1223251d-7186-4aa3-8d6a-f19b3f248512","Type":"ContainerDied","Data":"6f598c51ab6b00be740b404712f1b796dd2890b2479dcbfb419971c8a29c2989"} Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.107984 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"64660a67-5194-45d4-a364-228db58cf2a9","Type":"ContainerStarted","Data":"87a6f7da303059b5522a0750ac976d61582986649756357e60cfe804f62a453a"} Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.604226 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.705633 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppzbt\" (UniqueName: \"kubernetes.io/projected/1223251d-7186-4aa3-8d6a-f19b3f248512-kube-api-access-ppzbt\") pod \"1223251d-7186-4aa3-8d6a-f19b3f248512\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.706141 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-config-data\") pod \"1223251d-7186-4aa3-8d6a-f19b3f248512\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.706621 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-combined-ca-bundle\") pod \"1223251d-7186-4aa3-8d6a-f19b3f248512\" (UID: \"1223251d-7186-4aa3-8d6a-f19b3f248512\") " Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.730286 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1223251d-7186-4aa3-8d6a-f19b3f248512-kube-api-access-ppzbt" (OuterVolumeSpecName: "kube-api-access-ppzbt") pod "1223251d-7186-4aa3-8d6a-f19b3f248512" (UID: "1223251d-7186-4aa3-8d6a-f19b3f248512"). InnerVolumeSpecName "kube-api-access-ppzbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.759213 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-config-data" (OuterVolumeSpecName: "config-data") pod "1223251d-7186-4aa3-8d6a-f19b3f248512" (UID: "1223251d-7186-4aa3-8d6a-f19b3f248512"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.759620 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1223251d-7186-4aa3-8d6a-f19b3f248512" (UID: "1223251d-7186-4aa3-8d6a-f19b3f248512"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.811643 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppzbt\" (UniqueName: \"kubernetes.io/projected/1223251d-7186-4aa3-8d6a-f19b3f248512-kube-api-access-ppzbt\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.811690 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.811702 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1223251d-7186-4aa3-8d6a-f19b3f248512-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:07 crc kubenswrapper[4876]: I1205 08:38:07.844005 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e42626e-224c-4d7d-9fb9-5620eec12bd6" path="/var/lib/kubelet/pods/4e42626e-224c-4d7d-9fb9-5620eec12bd6/volumes" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.043258 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.121124 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerStarted","Data":"22e3f0467be0882ca230a6aa9bc8bb7912fcd0ec96e095588b54f8094222ab90"} Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.124550 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1223251d-7186-4aa3-8d6a-f19b3f248512","Type":"ContainerDied","Data":"1d16eeb0818d6f0ee1afcaf230ffa17b340df18d3d188fa88740aa6e4058a363"} Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.124585 4876 scope.go:117] "RemoveContainer" containerID="6f598c51ab6b00be740b404712f1b796dd2890b2479dcbfb419971c8a29c2989" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.124711 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.145450 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gpm57" event={"ID":"198b165f-0f97-4551-a939-73d07f958d01","Type":"ContainerStarted","Data":"fcc9274310b57ba2fa1cba5b1f467a6e7fe5d2ca5ed7ea581df54d9d31ce767c"} Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.152927 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.172155 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.172569 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" event={"ID":"14bc9681-6da3-4e8b-8a88-4940f7b359de","Type":"ContainerStarted","Data":"5a342e34f983f5092466e9864f0f8c3cf28865539e000d238e11d3aaf9f6ab15"} Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.173599 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.185503 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-gpm57" podStartSLOduration=2.185481536 podStartE2EDuration="2.185481536s" podCreationTimestamp="2025-12-05 08:38:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:08.162374044 +0000 UTC m=+1352.651038666" watchObservedRunningTime="2025-12-05 08:38:08.185481536 +0000 UTC m=+1352.674146158" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.204086 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:38:08 crc kubenswrapper[4876]: E1205 08:38:08.204487 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1223251d-7186-4aa3-8d6a-f19b3f248512" containerName="nova-cell0-conductor-conductor" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.204501 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1223251d-7186-4aa3-8d6a-f19b3f248512" containerName="nova-cell0-conductor-conductor" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.204699 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1223251d-7186-4aa3-8d6a-f19b3f248512" containerName="nova-cell0-conductor-conductor" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.205311 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.207296 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.223191 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.229211 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" podStartSLOduration=4.229189821 podStartE2EDuration="4.229189821s" podCreationTimestamp="2025-12-05 08:38:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:08.202424241 +0000 UTC m=+1352.691088853" watchObservedRunningTime="2025-12-05 08:38:08.229189821 +0000 UTC m=+1352.717854453" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.325959 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd8v6\" (UniqueName: \"kubernetes.io/projected/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-kube-api-access-zd8v6\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.326367 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.327523 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.431928 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.434104 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.434560 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd8v6\" (UniqueName: \"kubernetes.io/projected/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-kube-api-access-zd8v6\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.443590 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.451593 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.453736 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd8v6\" (UniqueName: \"kubernetes.io/projected/dc5cc0db-4ac5-4093-a0f0-abf202104b4f-kube-api-access-zd8v6\") pod \"nova-cell0-conductor-0\" (UID: \"dc5cc0db-4ac5-4093-a0f0-abf202104b4f\") " pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:08 crc kubenswrapper[4876]: I1205 08:38:08.524193 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:09 crc kubenswrapper[4876]: I1205 08:38:09.856278 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1223251d-7186-4aa3-8d6a-f19b3f248512" path="/var/lib/kubelet/pods/1223251d-7186-4aa3-8d6a-f19b3f248512/volumes" Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.096352 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 08:38:10 crc kubenswrapper[4876]: W1205 08:38:10.103518 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc5cc0db_4ac5_4093_a0f0_abf202104b4f.slice/crio-8b44f1c957ea7612740617214f6701ac702e025e8a847825b349876abcf3885a WatchSource:0}: Error finding container 8b44f1c957ea7612740617214f6701ac702e025e8a847825b349876abcf3885a: Status 404 returned error can't find the container with id 8b44f1c957ea7612740617214f6701ac702e025e8a847825b349876abcf3885a Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.202206 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"64660a67-5194-45d4-a364-228db58cf2a9","Type":"ContainerStarted","Data":"b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611"} Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.203688 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f","Type":"ContainerStarted","Data":"e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5"} Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.206424 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6eb8b37-115c-4d30-9358-c4773f3e1d9e","Type":"ContainerStarted","Data":"b7df567c5a926779acae253ecc2985052873376cafeab669d894bd3a1aceffbe"} Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.208658 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9291e5f9-2b2d-4785-bdb4-0e43c997f458","Type":"ContainerStarted","Data":"24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d"} Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.212539 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dc5cc0db-4ac5-4093-a0f0-abf202104b4f","Type":"ContainerStarted","Data":"8b44f1c957ea7612740617214f6701ac702e025e8a847825b349876abcf3885a"} Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.227424 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerStarted","Data":"2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54"} Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.228476 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.24476883 podStartE2EDuration="6.22846037s" podCreationTimestamp="2025-12-05 08:38:04 +0000 UTC" firstStartedPulling="2025-12-05 08:38:05.642852317 +0000 UTC m=+1350.131516939" lastFinishedPulling="2025-12-05 08:38:09.626543857 +0000 UTC m=+1354.115208479" observedRunningTime="2025-12-05 08:38:10.216152319 +0000 UTC m=+1354.704816941" watchObservedRunningTime="2025-12-05 08:38:10.22846037 +0000 UTC m=+1354.717124992" Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.236362 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.820622462 podStartE2EDuration="6.236342752s" podCreationTimestamp="2025-12-05 08:38:04 +0000 UTC" firstStartedPulling="2025-12-05 08:38:06.233765254 +0000 UTC m=+1350.722429876" lastFinishedPulling="2025-12-05 08:38:09.649485544 +0000 UTC m=+1354.138150166" observedRunningTime="2025-12-05 08:38:10.233366522 +0000 UTC m=+1354.722031144" watchObservedRunningTime="2025-12-05 08:38:10.236342752 +0000 UTC m=+1354.725007374" Dec 05 08:38:10 crc kubenswrapper[4876]: I1205 08:38:10.335530 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:11 crc kubenswrapper[4876]: I1205 08:38:11.265350 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dc5cc0db-4ac5-4093-a0f0-abf202104b4f","Type":"ContainerStarted","Data":"9fe485ad17c3057695be00353fdd52b7d8479d68703c8e1a461a5a2975734457"} Dec 05 08:38:11 crc kubenswrapper[4876]: I1205 08:38:11.265874 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:11 crc kubenswrapper[4876]: I1205 08:38:11.271796 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerStarted","Data":"9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25"} Dec 05 08:38:11 crc kubenswrapper[4876]: I1205 08:38:11.274225 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"64660a67-5194-45d4-a364-228db58cf2a9","Type":"ContainerStarted","Data":"f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a"} Dec 05 08:38:11 crc kubenswrapper[4876]: I1205 08:38:11.276825 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6eb8b37-115c-4d30-9358-c4773f3e1d9e","Type":"ContainerStarted","Data":"a8aec8fd350c5efe53e4242c252de2cb7bbd54414be4134135a0af1e76302ff7"} Dec 05 08:38:11 crc kubenswrapper[4876]: I1205 08:38:11.289854 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.289829755 podStartE2EDuration="3.289829755s" podCreationTimestamp="2025-12-05 08:38:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:11.280695969 +0000 UTC m=+1355.769360591" watchObservedRunningTime="2025-12-05 08:38:11.289829755 +0000 UTC m=+1355.778494377" Dec 05 08:38:11 crc kubenswrapper[4876]: I1205 08:38:11.304695 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.180127207 podStartE2EDuration="7.304671404s" podCreationTimestamp="2025-12-05 08:38:04 +0000 UTC" firstStartedPulling="2025-12-05 08:38:05.490691906 +0000 UTC m=+1349.979356528" lastFinishedPulling="2025-12-05 08:38:09.615236103 +0000 UTC m=+1354.103900725" observedRunningTime="2025-12-05 08:38:11.29820501 +0000 UTC m=+1355.786869632" watchObservedRunningTime="2025-12-05 08:38:11.304671404 +0000 UTC m=+1355.793336026" Dec 05 08:38:11 crc kubenswrapper[4876]: I1205 08:38:11.320158 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.796820666 podStartE2EDuration="7.32013993s" podCreationTimestamp="2025-12-05 08:38:04 +0000 UTC" firstStartedPulling="2025-12-05 08:38:06.131767161 +0000 UTC m=+1350.620431783" lastFinishedPulling="2025-12-05 08:38:09.655086425 +0000 UTC m=+1354.143751047" observedRunningTime="2025-12-05 08:38:11.313690476 +0000 UTC m=+1355.802355088" watchObservedRunningTime="2025-12-05 08:38:11.32013993 +0000 UTC m=+1355.808804552" Dec 05 08:38:12 crc kubenswrapper[4876]: I1205 08:38:12.287836 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerStarted","Data":"bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb"} Dec 05 08:38:13 crc kubenswrapper[4876]: I1205 08:38:13.302089 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerStarted","Data":"104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9"} Dec 05 08:38:13 crc kubenswrapper[4876]: I1205 08:38:13.303212 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:38:13 crc kubenswrapper[4876]: I1205 08:38:13.303083 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="proxy-httpd" containerID="cri-o://104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9" gracePeriod=30 Dec 05 08:38:13 crc kubenswrapper[4876]: I1205 08:38:13.302495 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="ceilometer-central-agent" containerID="cri-o://2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54" gracePeriod=30 Dec 05 08:38:13 crc kubenswrapper[4876]: I1205 08:38:13.303132 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="ceilometer-notification-agent" containerID="cri-o://9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25" gracePeriod=30 Dec 05 08:38:13 crc kubenswrapper[4876]: I1205 08:38:13.303116 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="sg-core" containerID="cri-o://bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb" gracePeriod=30 Dec 05 08:38:13 crc kubenswrapper[4876]: I1205 08:38:13.330746 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.48230031 podStartE2EDuration="7.330726383s" podCreationTimestamp="2025-12-05 08:38:06 +0000 UTC" firstStartedPulling="2025-12-05 08:38:07.133091532 +0000 UTC m=+1351.621756154" lastFinishedPulling="2025-12-05 08:38:12.981517595 +0000 UTC m=+1357.470182227" observedRunningTime="2025-12-05 08:38:13.324851955 +0000 UTC m=+1357.813516577" watchObservedRunningTime="2025-12-05 08:38:13.330726383 +0000 UTC m=+1357.819391005" Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.314664 4876 generic.go:334] "Generic (PLEG): container finished" podID="b6629f58-7171-4cd4-a018-6b9c40a8bfee" containerID="61ca35015ec9ca00cca71b6d6bd2a6cd1722b775213b9192131b407f5798c455" exitCode=0 Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.314726 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vhxdf" event={"ID":"b6629f58-7171-4cd4-a018-6b9c40a8bfee","Type":"ContainerDied","Data":"61ca35015ec9ca00cca71b6d6bd2a6cd1722b775213b9192131b407f5798c455"} Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.319516 4876 generic.go:334] "Generic (PLEG): container finished" podID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerID="bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb" exitCode=2 Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.319535 4876 generic.go:334] "Generic (PLEG): container finished" podID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerID="9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25" exitCode=0 Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.319551 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerDied","Data":"bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb"} Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.319567 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerDied","Data":"9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25"} Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.619047 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.619143 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.979619 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 08:38:14 crc kubenswrapper[4876]: I1205 08:38:14.979685 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.010862 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.158230 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.218592 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-tw8hl"] Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.219304 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" podUID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerName="dnsmasq-dns" containerID="cri-o://a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c" gracePeriod=10 Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.298738 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.299064 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.299087 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.299106 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.340461 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.347299 4876 generic.go:334] "Generic (PLEG): container finished" podID="198b165f-0f97-4551-a939-73d07f958d01" containerID="fcc9274310b57ba2fa1cba5b1f467a6e7fe5d2ca5ed7ea581df54d9d31ce767c" exitCode=0 Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.347760 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gpm57" event={"ID":"198b165f-0f97-4551-a939-73d07f958d01","Type":"ContainerDied","Data":"fcc9274310b57ba2fa1cba5b1f467a6e7fe5d2ca5ed7ea581df54d9d31ce767c"} Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.373705 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.386783 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.711097 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:15 crc kubenswrapper[4876]: I1205 08:38:15.711250 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.043339 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.057507 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216325 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-combined-ca-bundle\") pod \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216400 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-config\") pod \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216474 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-scripts\") pod \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216526 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxvmt\" (UniqueName: \"kubernetes.io/projected/bd36d5f7-6681-4eba-a08c-28357a52cf5e-kube-api-access-fxvmt\") pod \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216546 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-sb\") pod \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216655 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-config-data\") pod \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216693 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-swift-storage-0\") pod \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216716 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bnfb\" (UniqueName: \"kubernetes.io/projected/b6629f58-7171-4cd4-a018-6b9c40a8bfee-kube-api-access-7bnfb\") pod \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\" (UID: \"b6629f58-7171-4cd4-a018-6b9c40a8bfee\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216736 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-svc\") pod \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.216762 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-nb\") pod \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\" (UID: \"bd36d5f7-6681-4eba-a08c-28357a52cf5e\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.240160 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6629f58-7171-4cd4-a018-6b9c40a8bfee-kube-api-access-7bnfb" (OuterVolumeSpecName: "kube-api-access-7bnfb") pod "b6629f58-7171-4cd4-a018-6b9c40a8bfee" (UID: "b6629f58-7171-4cd4-a018-6b9c40a8bfee"). InnerVolumeSpecName "kube-api-access-7bnfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.254836 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-scripts" (OuterVolumeSpecName: "scripts") pod "b6629f58-7171-4cd4-a018-6b9c40a8bfee" (UID: "b6629f58-7171-4cd4-a018-6b9c40a8bfee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.254860 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd36d5f7-6681-4eba-a08c-28357a52cf5e-kube-api-access-fxvmt" (OuterVolumeSpecName: "kube-api-access-fxvmt") pod "bd36d5f7-6681-4eba-a08c-28357a52cf5e" (UID: "bd36d5f7-6681-4eba-a08c-28357a52cf5e"). InnerVolumeSpecName "kube-api-access-fxvmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.301013 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6629f58-7171-4cd4-a018-6b9c40a8bfee" (UID: "b6629f58-7171-4cd4-a018-6b9c40a8bfee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.307749 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-config-data" (OuterVolumeSpecName: "config-data") pod "b6629f58-7171-4cd4-a018-6b9c40a8bfee" (UID: "b6629f58-7171-4cd4-a018-6b9c40a8bfee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.318427 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.318462 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxvmt\" (UniqueName: \"kubernetes.io/projected/bd36d5f7-6681-4eba-a08c-28357a52cf5e-kube-api-access-fxvmt\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.318475 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.318483 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bnfb\" (UniqueName: \"kubernetes.io/projected/b6629f58-7171-4cd4-a018-6b9c40a8bfee-kube-api-access-7bnfb\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.318491 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6629f58-7171-4cd4-a018-6b9c40a8bfee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.339037 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bd36d5f7-6681-4eba-a08c-28357a52cf5e" (UID: "bd36d5f7-6681-4eba-a08c-28357a52cf5e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.342028 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bd36d5f7-6681-4eba-a08c-28357a52cf5e" (UID: "bd36d5f7-6681-4eba-a08c-28357a52cf5e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.350023 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd36d5f7-6681-4eba-a08c-28357a52cf5e" (UID: "bd36d5f7-6681-4eba-a08c-28357a52cf5e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.381156 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.0.191:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.381442 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.0.191:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.381694 4876 generic.go:334] "Generic (PLEG): container finished" podID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerID="a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c" exitCode=0 Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.381969 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.382437 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" event={"ID":"bd36d5f7-6681-4eba-a08c-28357a52cf5e","Type":"ContainerDied","Data":"a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c"} Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.382562 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" event={"ID":"bd36d5f7-6681-4eba-a08c-28357a52cf5e","Type":"ContainerDied","Data":"028cbac9b6ca7009a4811604d5e7f4dcfdad322e0f64810764422fef7110ccd4"} Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.382590 4876 scope.go:117] "RemoveContainer" containerID="a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.400019 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-config" (OuterVolumeSpecName: "config") pod "bd36d5f7-6681-4eba-a08c-28357a52cf5e" (UID: "bd36d5f7-6681-4eba-a08c-28357a52cf5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.408353 4876 generic.go:334] "Generic (PLEG): container finished" podID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerID="2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54" exitCode=0 Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.408417 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerDied","Data":"2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54"} Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.415221 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bd36d5f7-6681-4eba-a08c-28357a52cf5e" (UID: "bd36d5f7-6681-4eba-a08c-28357a52cf5e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.415287 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vhxdf" event={"ID":"b6629f58-7171-4cd4-a018-6b9c40a8bfee","Type":"ContainerDied","Data":"068fb74cc90c5981f8cbeffab4a1dfdc90ad7c82a883eddb99c2651cc2bb018d"} Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.415363 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="068fb74cc90c5981f8cbeffab4a1dfdc90ad7c82a883eddb99c2651cc2bb018d" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.415431 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vhxdf" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.420101 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.420132 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.420144 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.420156 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.420167 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd36d5f7-6681-4eba-a08c-28357a52cf5e-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.425064 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.528433 4876 scope.go:117] "RemoveContainer" containerID="e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.612086 4876 scope.go:117] "RemoveContainer" containerID="a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c" Dec 05 08:38:16 crc kubenswrapper[4876]: E1205 08:38:16.616494 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c\": container with ID starting with a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c not found: ID does not exist" containerID="a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.616658 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c"} err="failed to get container status \"a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c\": rpc error: code = NotFound desc = could not find container \"a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c\": container with ID starting with a9928b89857c4d950acf916eb5bda27ebe45fe9992daf125fd9cc9d65557483c not found: ID does not exist" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.616758 4876 scope.go:117] "RemoveContainer" containerID="e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3" Dec 05 08:38:16 crc kubenswrapper[4876]: E1205 08:38:16.619482 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3\": container with ID starting with e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3 not found: ID does not exist" containerID="e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.619534 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3"} err="failed to get container status \"e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3\": rpc error: code = NotFound desc = could not find container \"e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3\": container with ID starting with e6f9befa96832cd945d6bf43759e9ca8537d50ae6017c9b2562ba3899e8f0cf3 not found: ID does not exist" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.728737 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-tw8hl"] Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.734053 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-tw8hl"] Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.846789 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.929569 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-config-data\") pod \"198b165f-0f97-4551-a939-73d07f958d01\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.929712 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-combined-ca-bundle\") pod \"198b165f-0f97-4551-a939-73d07f958d01\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.929755 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-scripts\") pod \"198b165f-0f97-4551-a939-73d07f958d01\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.929857 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skz2r\" (UniqueName: \"kubernetes.io/projected/198b165f-0f97-4551-a939-73d07f958d01-kube-api-access-skz2r\") pod \"198b165f-0f97-4551-a939-73d07f958d01\" (UID: \"198b165f-0f97-4551-a939-73d07f958d01\") " Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.933530 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-scripts" (OuterVolumeSpecName: "scripts") pod "198b165f-0f97-4551-a939-73d07f958d01" (UID: "198b165f-0f97-4551-a939-73d07f958d01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.933793 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/198b165f-0f97-4551-a939-73d07f958d01-kube-api-access-skz2r" (OuterVolumeSpecName: "kube-api-access-skz2r") pod "198b165f-0f97-4551-a939-73d07f958d01" (UID: "198b165f-0f97-4551-a939-73d07f958d01"). InnerVolumeSpecName "kube-api-access-skz2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.957263 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-config-data" (OuterVolumeSpecName: "config-data") pod "198b165f-0f97-4551-a939-73d07f958d01" (UID: "198b165f-0f97-4551-a939-73d07f958d01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:16 crc kubenswrapper[4876]: I1205 08:38:16.963471 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "198b165f-0f97-4551-a939-73d07f958d01" (UID: "198b165f-0f97-4551-a939-73d07f958d01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.031784 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.031830 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.031843 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/198b165f-0f97-4551-a939-73d07f958d01-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.031854 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skz2r\" (UniqueName: \"kubernetes.io/projected/198b165f-0f97-4551-a939-73d07f958d01-kube-api-access-skz2r\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.436782 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gpm57" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.437060 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gpm57" event={"ID":"198b165f-0f97-4551-a939-73d07f958d01","Type":"ContainerDied","Data":"b3980706c4b98c83553537121c7a43e78596f47ff2222e49cae387ad8bb3af62"} Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.437088 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3980706c4b98c83553537121c7a43e78596f47ff2222e49cae387ad8bb3af62" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.481978 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:38:17 crc kubenswrapper[4876]: E1205 08:38:17.483065 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerName="dnsmasq-dns" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.483097 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerName="dnsmasq-dns" Dec 05 08:38:17 crc kubenswrapper[4876]: E1205 08:38:17.483152 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6629f58-7171-4cd4-a018-6b9c40a8bfee" containerName="nova-manage" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.483162 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6629f58-7171-4cd4-a018-6b9c40a8bfee" containerName="nova-manage" Dec 05 08:38:17 crc kubenswrapper[4876]: E1205 08:38:17.483185 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="198b165f-0f97-4551-a939-73d07f958d01" containerName="nova-cell1-conductor-db-sync" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.483194 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="198b165f-0f97-4551-a939-73d07f958d01" containerName="nova-cell1-conductor-db-sync" Dec 05 08:38:17 crc kubenswrapper[4876]: E1205 08:38:17.483248 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerName="init" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.483258 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerName="init" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.483749 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerName="dnsmasq-dns" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.483781 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="198b165f-0f97-4551-a939-73d07f958d01" containerName="nova-cell1-conductor-db-sync" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.483816 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6629f58-7171-4cd4-a018-6b9c40a8bfee" containerName="nova-manage" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.484889 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.493684 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.504544 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.539473 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0643cc1-c10a-4da7-a3b8-09063ec41b80-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.539551 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0643cc1-c10a-4da7-a3b8-09063ec41b80-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.539576 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx62t\" (UniqueName: \"kubernetes.io/projected/b0643cc1-c10a-4da7-a3b8-09063ec41b80-kube-api-access-nx62t\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.641837 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0643cc1-c10a-4da7-a3b8-09063ec41b80-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.641916 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0643cc1-c10a-4da7-a3b8-09063ec41b80-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.641946 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx62t\" (UniqueName: \"kubernetes.io/projected/b0643cc1-c10a-4da7-a3b8-09063ec41b80-kube-api-access-nx62t\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.648874 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0643cc1-c10a-4da7-a3b8-09063ec41b80-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.660605 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0643cc1-c10a-4da7-a3b8-09063ec41b80-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.662698 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx62t\" (UniqueName: \"kubernetes.io/projected/b0643cc1-c10a-4da7-a3b8-09063ec41b80-kube-api-access-nx62t\") pod \"nova-cell1-conductor-0\" (UID: \"b0643cc1-c10a-4da7-a3b8-09063ec41b80\") " pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.814231 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:17 crc kubenswrapper[4876]: I1205 08:38:17.841639 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" path="/var/lib/kubelet/pods/bd36d5f7-6681-4eba-a08c-28357a52cf5e/volumes" Dec 05 08:38:18 crc kubenswrapper[4876]: I1205 08:38:18.285768 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 08:38:18 crc kubenswrapper[4876]: W1205 08:38:18.289274 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0643cc1_c10a_4da7_a3b8_09063ec41b80.slice/crio-6a8202a1653fa50a3d75bef72c6a513df6238f59895e0ae1a9435f2972f26600 WatchSource:0}: Error finding container 6a8202a1653fa50a3d75bef72c6a513df6238f59895e0ae1a9435f2972f26600: Status 404 returned error can't find the container with id 6a8202a1653fa50a3d75bef72c6a513df6238f59895e0ae1a9435f2972f26600 Dec 05 08:38:18 crc kubenswrapper[4876]: I1205 08:38:18.450655 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b0643cc1-c10a-4da7-a3b8-09063ec41b80","Type":"ContainerStarted","Data":"6a8202a1653fa50a3d75bef72c6a513df6238f59895e0ae1a9435f2972f26600"} Dec 05 08:38:18 crc kubenswrapper[4876]: I1205 08:38:18.554122 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.025109 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.025403 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-log" containerID="cri-o://b7df567c5a926779acae253ecc2985052873376cafeab669d894bd3a1aceffbe" gracePeriod=30 Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.025477 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-api" containerID="cri-o://a8aec8fd350c5efe53e4242c252de2cb7bbd54414be4134135a0af1e76302ff7" gracePeriod=30 Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.045479 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.045724 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" containerName="nova-scheduler-scheduler" containerID="cri-o://e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5" gracePeriod=30 Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.099990 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.100185 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="9291e5f9-2b2d-4785-bdb4-0e43c997f458" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d" gracePeriod=30 Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.285922 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.286133 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-log" containerID="cri-o://b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611" gracePeriod=30 Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.286536 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-metadata" containerID="cri-o://f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a" gracePeriod=30 Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.466364 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"b0643cc1-c10a-4da7-a3b8-09063ec41b80","Type":"ContainerStarted","Data":"cab3f3a4d7bde12f220b227faabd67d77ec9309bcbefced0528898d45350b342"} Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.466441 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.471242 4876 generic.go:334] "Generic (PLEG): container finished" podID="64660a67-5194-45d4-a364-228db58cf2a9" containerID="b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611" exitCode=143 Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.471321 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"64660a67-5194-45d4-a364-228db58cf2a9","Type":"ContainerDied","Data":"b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611"} Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.473497 4876 generic.go:334] "Generic (PLEG): container finished" podID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerID="b7df567c5a926779acae253ecc2985052873376cafeab669d894bd3a1aceffbe" exitCode=143 Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.473536 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6eb8b37-115c-4d30-9358-c4773f3e1d9e","Type":"ContainerDied","Data":"b7df567c5a926779acae253ecc2985052873376cafeab669d894bd3a1aceffbe"} Dec 05 08:38:19 crc kubenswrapper[4876]: I1205 08:38:19.488420 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.488401504 podStartE2EDuration="2.488401504s" podCreationTimestamp="2025-12-05 08:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:19.481270773 +0000 UTC m=+1363.969935395" watchObservedRunningTime="2025-12-05 08:38:19.488401504 +0000 UTC m=+1363.977066126" Dec 05 08:38:19 crc kubenswrapper[4876]: E1205 08:38:19.981394 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:38:19 crc kubenswrapper[4876]: E1205 08:38:19.983244 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:38:19 crc kubenswrapper[4876]: E1205 08:38:19.986992 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:38:19 crc kubenswrapper[4876]: E1205 08:38:19.987044 4876 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" containerName="nova-scheduler-scheduler" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.010911 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.187685 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-combined-ca-bundle\") pod \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.187773 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-config-data\") pod \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.188217 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntg69\" (UniqueName: \"kubernetes.io/projected/9291e5f9-2b2d-4785-bdb4-0e43c997f458-kube-api-access-ntg69\") pod \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\" (UID: \"9291e5f9-2b2d-4785-bdb4-0e43c997f458\") " Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.192118 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9291e5f9-2b2d-4785-bdb4-0e43c997f458-kube-api-access-ntg69" (OuterVolumeSpecName: "kube-api-access-ntg69") pod "9291e5f9-2b2d-4785-bdb4-0e43c997f458" (UID: "9291e5f9-2b2d-4785-bdb4-0e43c997f458"). InnerVolumeSpecName "kube-api-access-ntg69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.215514 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-config-data" (OuterVolumeSpecName: "config-data") pod "9291e5f9-2b2d-4785-bdb4-0e43c997f458" (UID: "9291e5f9-2b2d-4785-bdb4-0e43c997f458"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.219574 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9291e5f9-2b2d-4785-bdb4-0e43c997f458" (UID: "9291e5f9-2b2d-4785-bdb4-0e43c997f458"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.292236 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntg69\" (UniqueName: \"kubernetes.io/projected/9291e5f9-2b2d-4785-bdb4-0e43c997f458-kube-api-access-ntg69\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.292281 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.292297 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9291e5f9-2b2d-4785-bdb4-0e43c997f458-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.486505 4876 generic.go:334] "Generic (PLEG): container finished" podID="9291e5f9-2b2d-4785-bdb4-0e43c997f458" containerID="24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d" exitCode=0 Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.486607 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.486662 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9291e5f9-2b2d-4785-bdb4-0e43c997f458","Type":"ContainerDied","Data":"24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d"} Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.486697 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9291e5f9-2b2d-4785-bdb4-0e43c997f458","Type":"ContainerDied","Data":"da55e836d4b01c08f93378809f388aba24cee1751e65419913f0b26b01fd3a6f"} Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.486717 4876 scope.go:117] "RemoveContainer" containerID="24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.524280 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.524974 4876 scope.go:117] "RemoveContainer" containerID="24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d" Dec 05 08:38:20 crc kubenswrapper[4876]: E1205 08:38:20.525518 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d\": container with ID starting with 24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d not found: ID does not exist" containerID="24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.525560 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d"} err="failed to get container status \"24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d\": rpc error: code = NotFound desc = could not find container \"24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d\": container with ID starting with 24cf2e04b319bb40a0a7c43a31e1380991b805bf0efcac9e78c791c7a8bf825d not found: ID does not exist" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.534578 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.553185 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:20 crc kubenswrapper[4876]: E1205 08:38:20.553569 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9291e5f9-2b2d-4785-bdb4-0e43c997f458" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.553581 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9291e5f9-2b2d-4785-bdb4-0e43c997f458" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.553789 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9291e5f9-2b2d-4785-bdb4-0e43c997f458" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.554421 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.558412 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.558614 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.563532 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.566178 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.606415 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.606544 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfzwc\" (UniqueName: \"kubernetes.io/projected/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-kube-api-access-rfzwc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.607017 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.607120 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.607513 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.708813 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.709154 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.709327 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.709505 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.709585 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfzwc\" (UniqueName: \"kubernetes.io/projected/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-kube-api-access-rfzwc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.713388 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.713681 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.714429 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.727520 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.738466 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfzwc\" (UniqueName: \"kubernetes.io/projected/8ee1cf7e-9e7b-407a-ab48-d4abb9326b27-kube-api-access-rfzwc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.780577 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7d978555f9-tw8hl" podUID="bd36d5f7-6681-4eba-a08c-28357a52cf5e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.165:5353: i/o timeout" Dec 05 08:38:20 crc kubenswrapper[4876]: I1205 08:38:20.917541 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:21 crc kubenswrapper[4876]: W1205 08:38:21.400814 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ee1cf7e_9e7b_407a_ab48_d4abb9326b27.slice/crio-5acee603719798e99a89fc97834fe6baf29082354ae1cfa22f03b799f869e65e WatchSource:0}: Error finding container 5acee603719798e99a89fc97834fe6baf29082354ae1cfa22f03b799f869e65e: Status 404 returned error can't find the container with id 5acee603719798e99a89fc97834fe6baf29082354ae1cfa22f03b799f869e65e Dec 05 08:38:21 crc kubenswrapper[4876]: I1205 08:38:21.406183 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 08:38:21 crc kubenswrapper[4876]: I1205 08:38:21.499872 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27","Type":"ContainerStarted","Data":"5acee603719798e99a89fc97834fe6baf29082354ae1cfa22f03b799f869e65e"} Dec 05 08:38:21 crc kubenswrapper[4876]: I1205 08:38:21.842469 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9291e5f9-2b2d-4785-bdb4-0e43c997f458" path="/var/lib/kubelet/pods/9291e5f9-2b2d-4785-bdb4-0e43c997f458/volumes" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.072854 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.136671 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6nbq\" (UniqueName: \"kubernetes.io/projected/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-kube-api-access-x6nbq\") pod \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.136797 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-config-data\") pod \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.136842 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-combined-ca-bundle\") pod \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\" (UID: \"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f\") " Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.142857 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-kube-api-access-x6nbq" (OuterVolumeSpecName: "kube-api-access-x6nbq") pod "9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" (UID: "9e08ab8f-c4d6-4688-bd9a-6a5d9784797f"). InnerVolumeSpecName "kube-api-access-x6nbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.145792 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6nbq\" (UniqueName: \"kubernetes.io/projected/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-kube-api-access-x6nbq\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.172109 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" (UID: "9e08ab8f-c4d6-4688-bd9a-6a5d9784797f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.183085 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-config-data" (OuterVolumeSpecName: "config-data") pod "9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" (UID: "9e08ab8f-c4d6-4688-bd9a-6a5d9784797f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.247729 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.247953 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.392488 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-57ztd"] Dec 05 08:38:22 crc kubenswrapper[4876]: E1205 08:38:22.392871 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" containerName="nova-scheduler-scheduler" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.392888 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" containerName="nova-scheduler-scheduler" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.393211 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" containerName="nova-scheduler-scheduler" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.394689 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.451009 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khz4g\" (UniqueName: \"kubernetes.io/projected/30bee729-e5e2-41a6-8aff-44f3efc4866f-kube-api-access-khz4g\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.451090 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-utilities\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.451215 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-catalog-content\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.454926 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-57ztd"] Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.533127 4876 generic.go:334] "Generic (PLEG): container finished" podID="9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" containerID="e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5" exitCode=0 Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.533256 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.533643 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f","Type":"ContainerDied","Data":"e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5"} Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.533685 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9e08ab8f-c4d6-4688-bd9a-6a5d9784797f","Type":"ContainerDied","Data":"ab1a3d93d2c56e19f2f087cfac518c932abc66011ebf6276ee48592b0e0a44f8"} Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.533703 4876 scope.go:117] "RemoveContainer" containerID="e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.552017 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-utilities\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.552118 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-catalog-content\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.552182 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khz4g\" (UniqueName: \"kubernetes.io/projected/30bee729-e5e2-41a6-8aff-44f3efc4866f-kube-api-access-khz4g\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.553275 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-utilities\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.553532 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-catalog-content\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.553722 4876 generic.go:334] "Generic (PLEG): container finished" podID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerID="a8aec8fd350c5efe53e4242c252de2cb7bbd54414be4134135a0af1e76302ff7" exitCode=0 Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.553800 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6eb8b37-115c-4d30-9358-c4773f3e1d9e","Type":"ContainerDied","Data":"a8aec8fd350c5efe53e4242c252de2cb7bbd54414be4134135a0af1e76302ff7"} Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.573002 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khz4g\" (UniqueName: \"kubernetes.io/projected/30bee729-e5e2-41a6-8aff-44f3efc4866f-kube-api-access-khz4g\") pod \"community-operators-57ztd\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.576065 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8ee1cf7e-9e7b-407a-ab48-d4abb9326b27","Type":"ContainerStarted","Data":"4eb3d1622dc07022af334c6219af15fb4de5e02eb1f41bd9ecb55733e8755f0b"} Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.672746 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.677942 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.677920089 podStartE2EDuration="2.677920089s" podCreationTimestamp="2025-12-05 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:22.608486322 +0000 UTC m=+1367.097150944" watchObservedRunningTime="2025-12-05 08:38:22.677920089 +0000 UTC m=+1367.166584711" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.686269 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.689514 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.700059 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.712311 4876 scope.go:117] "RemoveContainer" containerID="e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5" Dec 05 08:38:22 crc kubenswrapper[4876]: E1205 08:38:22.713724 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5\": container with ID starting with e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5 not found: ID does not exist" containerID="e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.713762 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5"} err="failed to get container status \"e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5\": rpc error: code = NotFound desc = could not find container \"e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5\": container with ID starting with e6a05bbd9ba4823911d92f6c037233e1138eb093341140b857c5926c7a833cf5 not found: ID does not exist" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.743954 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:22 crc kubenswrapper[4876]: E1205 08:38:22.744483 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-api" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.744503 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-api" Dec 05 08:38:22 crc kubenswrapper[4876]: E1205 08:38:22.744569 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-log" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.744580 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-log" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.744853 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-api" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.744892 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" containerName="nova-api-log" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.745737 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.750667 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.755951 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.759845 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-config-data\") pod \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.760006 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-logs\") pod \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.760088 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-combined-ca-bundle\") pod \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.760154 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdlz4\" (UniqueName: \"kubernetes.io/projected/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-kube-api-access-bdlz4\") pod \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\" (UID: \"f6eb8b37-115c-4d30-9358-c4773f3e1d9e\") " Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.760430 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqjw8\" (UniqueName: \"kubernetes.io/projected/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-kube-api-access-sqjw8\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.760453 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-logs" (OuterVolumeSpecName: "logs") pod "f6eb8b37-115c-4d30-9358-c4773f3e1d9e" (UID: "f6eb8b37-115c-4d30-9358-c4773f3e1d9e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.760802 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.761057 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-config-data\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.761300 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.763649 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-kube-api-access-bdlz4" (OuterVolumeSpecName: "kube-api-access-bdlz4") pod "f6eb8b37-115c-4d30-9358-c4773f3e1d9e" (UID: "f6eb8b37-115c-4d30-9358-c4773f3e1d9e"). InnerVolumeSpecName "kube-api-access-bdlz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.829647 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-config-data" (OuterVolumeSpecName: "config-data") pod "f6eb8b37-115c-4d30-9358-c4773f3e1d9e" (UID: "f6eb8b37-115c-4d30-9358-c4773f3e1d9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.869615 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-config-data\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.869675 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqjw8\" (UniqueName: \"kubernetes.io/projected/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-kube-api-access-sqjw8\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.869803 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.869875 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdlz4\" (UniqueName: \"kubernetes.io/projected/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-kube-api-access-bdlz4\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.869890 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.874517 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-config-data\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.877500 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.887840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6eb8b37-115c-4d30-9358-c4773f3e1d9e" (UID: "f6eb8b37-115c-4d30-9358-c4773f3e1d9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.904410 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqjw8\" (UniqueName: \"kubernetes.io/projected/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-kube-api-access-sqjw8\") pod \"nova-scheduler-0\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:22 crc kubenswrapper[4876]: I1205 08:38:22.975029 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6eb8b37-115c-4d30-9358-c4773f3e1d9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:22.998417 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.076771 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-config-data\") pod \"64660a67-5194-45d4-a364-228db58cf2a9\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.076876 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr5sb\" (UniqueName: \"kubernetes.io/projected/64660a67-5194-45d4-a364-228db58cf2a9-kube-api-access-mr5sb\") pod \"64660a67-5194-45d4-a364-228db58cf2a9\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.076922 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-combined-ca-bundle\") pod \"64660a67-5194-45d4-a364-228db58cf2a9\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.076946 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64660a67-5194-45d4-a364-228db58cf2a9-logs\") pod \"64660a67-5194-45d4-a364-228db58cf2a9\" (UID: \"64660a67-5194-45d4-a364-228db58cf2a9\") " Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.077530 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64660a67-5194-45d4-a364-228db58cf2a9-logs" (OuterVolumeSpecName: "logs") pod "64660a67-5194-45d4-a364-228db58cf2a9" (UID: "64660a67-5194-45d4-a364-228db58cf2a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.083243 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64660a67-5194-45d4-a364-228db58cf2a9-kube-api-access-mr5sb" (OuterVolumeSpecName: "kube-api-access-mr5sb") pod "64660a67-5194-45d4-a364-228db58cf2a9" (UID: "64660a67-5194-45d4-a364-228db58cf2a9"). InnerVolumeSpecName "kube-api-access-mr5sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.119212 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.137003 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64660a67-5194-45d4-a364-228db58cf2a9" (UID: "64660a67-5194-45d4-a364-228db58cf2a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.150166 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-config-data" (OuterVolumeSpecName: "config-data") pod "64660a67-5194-45d4-a364-228db58cf2a9" (UID: "64660a67-5194-45d4-a364-228db58cf2a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.184874 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.184915 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr5sb\" (UniqueName: \"kubernetes.io/projected/64660a67-5194-45d4-a364-228db58cf2a9-kube-api-access-mr5sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.184927 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64660a67-5194-45d4-a364-228db58cf2a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.184937 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64660a67-5194-45d4-a364-228db58cf2a9-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.410046 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-57ztd"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.589040 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57ztd" event={"ID":"30bee729-e5e2-41a6-8aff-44f3efc4866f","Type":"ContainerStarted","Data":"3aed6c853ac3408eff98cf52797bc1aa421dc7e1dd09a7148254ff6a2dd2c498"} Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.596871 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.596857 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f6eb8b37-115c-4d30-9358-c4773f3e1d9e","Type":"ContainerDied","Data":"15fd688450ef0b4adaf229d73d34cda3f34f51191a4be635ce608e71c084f7fa"} Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.597043 4876 scope.go:117] "RemoveContainer" containerID="a8aec8fd350c5efe53e4242c252de2cb7bbd54414be4134135a0af1e76302ff7" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.599736 4876 generic.go:334] "Generic (PLEG): container finished" podID="64660a67-5194-45d4-a364-228db58cf2a9" containerID="f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a" exitCode=0 Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.599845 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"64660a67-5194-45d4-a364-228db58cf2a9","Type":"ContainerDied","Data":"f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a"} Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.599870 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"64660a67-5194-45d4-a364-228db58cf2a9","Type":"ContainerDied","Data":"87a6f7da303059b5522a0750ac976d61582986649756357e60cfe804f62a453a"} Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.600002 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.623191 4876 scope.go:117] "RemoveContainer" containerID="b7df567c5a926779acae253ecc2985052873376cafeab669d894bd3a1aceffbe" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.638116 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.649346 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.671466 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: E1205 08:38:23.672235 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-metadata" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.672251 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-metadata" Dec 05 08:38:23 crc kubenswrapper[4876]: E1205 08:38:23.672287 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-log" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.672295 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-log" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.672480 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-log" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.672503 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="64660a67-5194-45d4-a364-228db58cf2a9" containerName="nova-metadata-metadata" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.673500 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.676573 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.676797 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.697370 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qrsn\" (UniqueName: \"kubernetes.io/projected/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-kube-api-access-4qrsn\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.697445 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-logs\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.697825 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-config-data\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.698023 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.698098 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.708953 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.709800 4876 scope.go:117] "RemoveContainer" containerID="f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.728719 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.753483 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.776967 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.778609 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.784303 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.788419 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.799783 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.799842 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.799884 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.799954 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qrsn\" (UniqueName: \"kubernetes.io/projected/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-kube-api-access-4qrsn\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.799973 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-logs\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.800003 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4e1018e-557a-47f3-bf3a-c2d62d83eece-logs\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.800022 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-config-data\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.800074 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhpq4\" (UniqueName: \"kubernetes.io/projected/a4e1018e-557a-47f3-bf3a-c2d62d83eece-kube-api-access-dhpq4\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.800104 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-config-data\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.804214 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-config-data\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.804428 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-logs\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.804490 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.807876 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.807960 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.825690 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qrsn\" (UniqueName: \"kubernetes.io/projected/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-kube-api-access-4qrsn\") pod \"nova-metadata-0\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " pod="openstack/nova-metadata-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.825850 4876 scope.go:117] "RemoveContainer" containerID="b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.838124 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64660a67-5194-45d4-a364-228db58cf2a9" path="/var/lib/kubelet/pods/64660a67-5194-45d4-a364-228db58cf2a9/volumes" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.839435 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e08ab8f-c4d6-4688-bd9a-6a5d9784797f" path="/var/lib/kubelet/pods/9e08ab8f-c4d6-4688-bd9a-6a5d9784797f/volumes" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.840262 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6eb8b37-115c-4d30-9358-c4773f3e1d9e" path="/var/lib/kubelet/pods/f6eb8b37-115c-4d30-9358-c4773f3e1d9e/volumes" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.861241 4876 scope.go:117] "RemoveContainer" containerID="f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a" Dec 05 08:38:23 crc kubenswrapper[4876]: E1205 08:38:23.861648 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a\": container with ID starting with f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a not found: ID does not exist" containerID="f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.861682 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a"} err="failed to get container status \"f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a\": rpc error: code = NotFound desc = could not find container \"f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a\": container with ID starting with f980e5bfc198b0f4d6b49dae1f4bb99740acc5548e07fd9415b6aea898e9714a not found: ID does not exist" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.861705 4876 scope.go:117] "RemoveContainer" containerID="b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611" Dec 05 08:38:23 crc kubenswrapper[4876]: E1205 08:38:23.862036 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611\": container with ID starting with b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611 not found: ID does not exist" containerID="b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.862061 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611"} err="failed to get container status \"b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611\": rpc error: code = NotFound desc = could not find container \"b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611\": container with ID starting with b1c805b76eb3738f19ffea0f9f2a7d019d4a9443697fc803c285eca222242611 not found: ID does not exist" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.901854 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhpq4\" (UniqueName: \"kubernetes.io/projected/a4e1018e-557a-47f3-bf3a-c2d62d83eece-kube-api-access-dhpq4\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.902016 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.902067 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4e1018e-557a-47f3-bf3a-c2d62d83eece-logs\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.902085 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-config-data\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.902553 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4e1018e-557a-47f3-bf3a-c2d62d83eece-logs\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.907278 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-config-data\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.910596 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:23 crc kubenswrapper[4876]: I1205 08:38:23.919680 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhpq4\" (UniqueName: \"kubernetes.io/projected/a4e1018e-557a-47f3-bf3a-c2d62d83eece-kube-api-access-dhpq4\") pod \"nova-api-0\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " pod="openstack/nova-api-0" Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.011368 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.180958 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.478206 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:24 crc kubenswrapper[4876]: W1205 08:38:24.492482 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1a13ae8_9c3c_43d6_af8e_cf13c3832166.slice/crio-71786f68c13dcfd6bea4de96d9c2a34b2145733180338aa1ce8abf0dc4e3823b WatchSource:0}: Error finding container 71786f68c13dcfd6bea4de96d9c2a34b2145733180338aa1ce8abf0dc4e3823b: Status 404 returned error can't find the container with id 71786f68c13dcfd6bea4de96d9c2a34b2145733180338aa1ce8abf0dc4e3823b Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.505952 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.615124 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4e1018e-557a-47f3-bf3a-c2d62d83eece","Type":"ContainerStarted","Data":"6861dcc4778a4e95e625753095f20d51aa72b0a6c4c6d7c0ec0d1f05d3f205e3"} Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.617345 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1a13ae8-9c3c-43d6-af8e-cf13c3832166","Type":"ContainerStarted","Data":"71786f68c13dcfd6bea4de96d9c2a34b2145733180338aa1ce8abf0dc4e3823b"} Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.621653 4876 generic.go:334] "Generic (PLEG): container finished" podID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerID="72d7deb181a1a661a588d9fe39fa19929159db7e6efb667c48f679ba0d70b6bb" exitCode=0 Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.621680 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57ztd" event={"ID":"30bee729-e5e2-41a6-8aff-44f3efc4866f","Type":"ContainerDied","Data":"72d7deb181a1a661a588d9fe39fa19929159db7e6efb667c48f679ba0d70b6bb"} Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.624614 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad","Type":"ContainerStarted","Data":"b7156e3de784c88fc1a2fac9ecfd83d657be9d95e8cc28801d49c86222436e94"} Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.624638 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad","Type":"ContainerStarted","Data":"f3b4a2e05f6ea3ab1959021c380d1ed3fe350a0a46f86d78be943c2d210634fb"} Dec 05 08:38:24 crc kubenswrapper[4876]: I1205 08:38:24.668027 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.668009164 podStartE2EDuration="2.668009164s" podCreationTimestamp="2025-12-05 08:38:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:24.660825102 +0000 UTC m=+1369.149489734" watchObservedRunningTime="2025-12-05 08:38:24.668009164 +0000 UTC m=+1369.156673786" Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.657037 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1a13ae8-9c3c-43d6-af8e-cf13c3832166","Type":"ContainerStarted","Data":"02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f"} Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.657086 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1a13ae8-9c3c-43d6-af8e-cf13c3832166","Type":"ContainerStarted","Data":"0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d"} Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.659849 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57ztd" event={"ID":"30bee729-e5e2-41a6-8aff-44f3efc4866f","Type":"ContainerDied","Data":"e2fb0baf57f5c55afa13ef8b7de60818ac9405b9d0da1b18a305012f545f444c"} Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.660465 4876 generic.go:334] "Generic (PLEG): container finished" podID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerID="e2fb0baf57f5c55afa13ef8b7de60818ac9405b9d0da1b18a305012f545f444c" exitCode=0 Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.665067 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4e1018e-557a-47f3-bf3a-c2d62d83eece","Type":"ContainerStarted","Data":"d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8"} Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.665121 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4e1018e-557a-47f3-bf3a-c2d62d83eece","Type":"ContainerStarted","Data":"8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf"} Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.693879 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.693857729 podStartE2EDuration="2.693857729s" podCreationTimestamp="2025-12-05 08:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:25.676111486 +0000 UTC m=+1370.164776168" watchObservedRunningTime="2025-12-05 08:38:25.693857729 +0000 UTC m=+1370.182522361" Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.721309 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.721288703 podStartE2EDuration="2.721288703s" podCreationTimestamp="2025-12-05 08:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:25.718431416 +0000 UTC m=+1370.207096048" watchObservedRunningTime="2025-12-05 08:38:25.721288703 +0000 UTC m=+1370.209953335" Dec 05 08:38:25 crc kubenswrapper[4876]: I1205 08:38:25.917891 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:26 crc kubenswrapper[4876]: I1205 08:38:26.676047 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57ztd" event={"ID":"30bee729-e5e2-41a6-8aff-44f3efc4866f","Type":"ContainerStarted","Data":"3e449e8c7ad5493b1493d68a00d78745cda67217c2160b3889dfa1b9047381ac"} Dec 05 08:38:26 crc kubenswrapper[4876]: I1205 08:38:26.698240 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-57ztd" podStartSLOduration=3.244226441 podStartE2EDuration="4.698222665s" podCreationTimestamp="2025-12-05 08:38:22 +0000 UTC" firstStartedPulling="2025-12-05 08:38:24.623787326 +0000 UTC m=+1369.112451938" lastFinishedPulling="2025-12-05 08:38:26.07778354 +0000 UTC m=+1370.566448162" observedRunningTime="2025-12-05 08:38:26.692776457 +0000 UTC m=+1371.181441109" watchObservedRunningTime="2025-12-05 08:38:26.698222665 +0000 UTC m=+1371.186887287" Dec 05 08:38:27 crc kubenswrapper[4876]: I1205 08:38:27.842998 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 08:38:28 crc kubenswrapper[4876]: I1205 08:38:28.119854 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 08:38:29 crc kubenswrapper[4876]: I1205 08:38:29.011921 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:38:29 crc kubenswrapper[4876]: I1205 08:38:29.011987 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:38:30 crc kubenswrapper[4876]: I1205 08:38:30.918267 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:30 crc kubenswrapper[4876]: I1205 08:38:30.938619 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:31 crc kubenswrapper[4876]: I1205 08:38:31.766460 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 08:38:31 crc kubenswrapper[4876]: I1205 08:38:31.908226 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-n696k"] Dec 05 08:38:31 crc kubenswrapper[4876]: I1205 08:38:31.910012 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:31 crc kubenswrapper[4876]: I1205 08:38:31.912221 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 05 08:38:31 crc kubenswrapper[4876]: I1205 08:38:31.917234 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 05 08:38:31 crc kubenswrapper[4876]: I1205 08:38:31.937220 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-n696k"] Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.004732 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg2xk\" (UniqueName: \"kubernetes.io/projected/36f2a90b-0c85-478b-83e5-17ff10e01c94-kube-api-access-xg2xk\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.004879 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-config-data\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.005076 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-scripts\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.005118 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.106768 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg2xk\" (UniqueName: \"kubernetes.io/projected/36f2a90b-0c85-478b-83e5-17ff10e01c94-kube-api-access-xg2xk\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.106863 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-config-data\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.107828 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-scripts\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.107865 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.112608 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-scripts\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.114194 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-config-data\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.125494 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg2xk\" (UniqueName: \"kubernetes.io/projected/36f2a90b-0c85-478b-83e5-17ff10e01c94-kube-api-access-xg2xk\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.144997 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-n696k\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.231289 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.673841 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.674255 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.719931 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-n696k"] Dec 05 08:38:32 crc kubenswrapper[4876]: W1205 08:38:32.721502 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36f2a90b_0c85_478b_83e5_17ff10e01c94.slice/crio-aafc6f43632f136e52c8ea982fab9a9fae34b7a57ddb08e5fc75403595acf1f3 WatchSource:0}: Error finding container aafc6f43632f136e52c8ea982fab9a9fae34b7a57ddb08e5fc75403595acf1f3: Status 404 returned error can't find the container with id aafc6f43632f136e52c8ea982fab9a9fae34b7a57ddb08e5fc75403595acf1f3 Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.763659 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-n696k" event={"ID":"36f2a90b-0c85-478b-83e5-17ff10e01c94","Type":"ContainerStarted","Data":"aafc6f43632f136e52c8ea982fab9a9fae34b7a57ddb08e5fc75403595acf1f3"} Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.768257 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:32 crc kubenswrapper[4876]: I1205 08:38:32.844423 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:33 crc kubenswrapper[4876]: I1205 08:38:33.120875 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 08:38:33 crc kubenswrapper[4876]: I1205 08:38:33.157553 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 08:38:33 crc kubenswrapper[4876]: I1205 08:38:33.773514 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-n696k" event={"ID":"36f2a90b-0c85-478b-83e5-17ff10e01c94","Type":"ContainerStarted","Data":"3eabe4487bdee6ddd8756e2c4aec68ac32a724d478587a901ac484e08b44507e"} Dec 05 08:38:33 crc kubenswrapper[4876]: I1205 08:38:33.794108 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-n696k" podStartSLOduration=2.794090915 podStartE2EDuration="2.794090915s" podCreationTimestamp="2025-12-05 08:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:33.785643327 +0000 UTC m=+1378.274307949" watchObservedRunningTime="2025-12-05 08:38:33.794090915 +0000 UTC m=+1378.282755537" Dec 05 08:38:33 crc kubenswrapper[4876]: I1205 08:38:33.807289 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 08:38:34 crc kubenswrapper[4876]: I1205 08:38:34.012547 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:38:34 crc kubenswrapper[4876]: I1205 08:38:34.012611 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:38:34 crc kubenswrapper[4876]: I1205 08:38:34.181840 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:38:34 crc kubenswrapper[4876]: I1205 08:38:34.181919 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:38:35 crc kubenswrapper[4876]: I1205 08:38:35.027172 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:35 crc kubenswrapper[4876]: I1205 08:38:35.027225 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:35 crc kubenswrapper[4876]: I1205 08:38:35.266116 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:35 crc kubenswrapper[4876]: I1205 08:38:35.266496 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:35 crc kubenswrapper[4876]: I1205 08:38:35.386310 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-57ztd"] Dec 05 08:38:35 crc kubenswrapper[4876]: I1205 08:38:35.386539 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-57ztd" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerName="registry-server" containerID="cri-o://3e449e8c7ad5493b1493d68a00d78745cda67217c2160b3889dfa1b9047381ac" gracePeriod=2 Dec 05 08:38:35 crc kubenswrapper[4876]: I1205 08:38:35.814399 4876 generic.go:334] "Generic (PLEG): container finished" podID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerID="3e449e8c7ad5493b1493d68a00d78745cda67217c2160b3889dfa1b9047381ac" exitCode=0 Dec 05 08:38:35 crc kubenswrapper[4876]: I1205 08:38:35.814437 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57ztd" event={"ID":"30bee729-e5e2-41a6-8aff-44f3efc4866f","Type":"ContainerDied","Data":"3e449e8c7ad5493b1493d68a00d78745cda67217c2160b3889dfa1b9047381ac"} Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.262610 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.421208 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-catalog-content\") pod \"30bee729-e5e2-41a6-8aff-44f3efc4866f\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.422443 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khz4g\" (UniqueName: \"kubernetes.io/projected/30bee729-e5e2-41a6-8aff-44f3efc4866f-kube-api-access-khz4g\") pod \"30bee729-e5e2-41a6-8aff-44f3efc4866f\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.422615 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-utilities\") pod \"30bee729-e5e2-41a6-8aff-44f3efc4866f\" (UID: \"30bee729-e5e2-41a6-8aff-44f3efc4866f\") " Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.423290 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-utilities" (OuterVolumeSpecName: "utilities") pod "30bee729-e5e2-41a6-8aff-44f3efc4866f" (UID: "30bee729-e5e2-41a6-8aff-44f3efc4866f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.434485 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30bee729-e5e2-41a6-8aff-44f3efc4866f-kube-api-access-khz4g" (OuterVolumeSpecName: "kube-api-access-khz4g") pod "30bee729-e5e2-41a6-8aff-44f3efc4866f" (UID: "30bee729-e5e2-41a6-8aff-44f3efc4866f"). InnerVolumeSpecName "kube-api-access-khz4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.469357 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.478275 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30bee729-e5e2-41a6-8aff-44f3efc4866f" (UID: "30bee729-e5e2-41a6-8aff-44f3efc4866f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.550334 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.550373 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khz4g\" (UniqueName: \"kubernetes.io/projected/30bee729-e5e2-41a6-8aff-44f3efc4866f-kube-api-access-khz4g\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.550385 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30bee729-e5e2-41a6-8aff-44f3efc4866f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.829542 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57ztd" event={"ID":"30bee729-e5e2-41a6-8aff-44f3efc4866f","Type":"ContainerDied","Data":"3aed6c853ac3408eff98cf52797bc1aa421dc7e1dd09a7148254ff6a2dd2c498"} Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.829614 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57ztd" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.829935 4876 scope.go:117] "RemoveContainer" containerID="3e449e8c7ad5493b1493d68a00d78745cda67217c2160b3889dfa1b9047381ac" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.870202 4876 scope.go:117] "RemoveContainer" containerID="e2fb0baf57f5c55afa13ef8b7de60818ac9405b9d0da1b18a305012f545f444c" Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.882329 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-57ztd"] Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.891167 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-57ztd"] Dec 05 08:38:36 crc kubenswrapper[4876]: I1205 08:38:36.903743 4876 scope.go:117] "RemoveContainer" containerID="72d7deb181a1a661a588d9fe39fa19929159db7e6efb667c48f679ba0d70b6bb" Dec 05 08:38:37 crc kubenswrapper[4876]: I1205 08:38:37.842283 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" path="/var/lib/kubelet/pods/30bee729-e5e2-41a6-8aff-44f3efc4866f/volumes" Dec 05 08:38:38 crc kubenswrapper[4876]: I1205 08:38:38.856889 4876 generic.go:334] "Generic (PLEG): container finished" podID="36f2a90b-0c85-478b-83e5-17ff10e01c94" containerID="3eabe4487bdee6ddd8756e2c4aec68ac32a724d478587a901ac484e08b44507e" exitCode=0 Dec 05 08:38:38 crc kubenswrapper[4876]: I1205 08:38:38.856963 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-n696k" event={"ID":"36f2a90b-0c85-478b-83e5-17ff10e01c94","Type":"ContainerDied","Data":"3eabe4487bdee6ddd8756e2c4aec68ac32a724d478587a901ac484e08b44507e"} Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.237237 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.434431 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-combined-ca-bundle\") pod \"36f2a90b-0c85-478b-83e5-17ff10e01c94\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.434469 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-config-data\") pod \"36f2a90b-0c85-478b-83e5-17ff10e01c94\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.434501 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-scripts\") pod \"36f2a90b-0c85-478b-83e5-17ff10e01c94\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.434581 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg2xk\" (UniqueName: \"kubernetes.io/projected/36f2a90b-0c85-478b-83e5-17ff10e01c94-kube-api-access-xg2xk\") pod \"36f2a90b-0c85-478b-83e5-17ff10e01c94\" (UID: \"36f2a90b-0c85-478b-83e5-17ff10e01c94\") " Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.440225 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-scripts" (OuterVolumeSpecName: "scripts") pod "36f2a90b-0c85-478b-83e5-17ff10e01c94" (UID: "36f2a90b-0c85-478b-83e5-17ff10e01c94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.440595 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36f2a90b-0c85-478b-83e5-17ff10e01c94-kube-api-access-xg2xk" (OuterVolumeSpecName: "kube-api-access-xg2xk") pod "36f2a90b-0c85-478b-83e5-17ff10e01c94" (UID: "36f2a90b-0c85-478b-83e5-17ff10e01c94"). InnerVolumeSpecName "kube-api-access-xg2xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.462283 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36f2a90b-0c85-478b-83e5-17ff10e01c94" (UID: "36f2a90b-0c85-478b-83e5-17ff10e01c94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.491857 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-config-data" (OuterVolumeSpecName: "config-data") pod "36f2a90b-0c85-478b-83e5-17ff10e01c94" (UID: "36f2a90b-0c85-478b-83e5-17ff10e01c94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.536609 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.536652 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.536663 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36f2a90b-0c85-478b-83e5-17ff10e01c94-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.536675 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg2xk\" (UniqueName: \"kubernetes.io/projected/36f2a90b-0c85-478b-83e5-17ff10e01c94-kube-api-access-xg2xk\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.875643 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-n696k" event={"ID":"36f2a90b-0c85-478b-83e5-17ff10e01c94","Type":"ContainerDied","Data":"aafc6f43632f136e52c8ea982fab9a9fae34b7a57ddb08e5fc75403595acf1f3"} Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.875685 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aafc6f43632f136e52c8ea982fab9a9fae34b7a57ddb08e5fc75403595acf1f3" Dec 05 08:38:40 crc kubenswrapper[4876]: I1205 08:38:40.875709 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-n696k" Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.109019 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.109761 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-log" containerID="cri-o://8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf" gracePeriod=30 Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.109810 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-api" containerID="cri-o://d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8" gracePeriod=30 Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.125701 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.125998 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" containerName="nova-scheduler-scheduler" containerID="cri-o://b7156e3de784c88fc1a2fac9ecfd83d657be9d95e8cc28801d49c86222436e94" gracePeriod=30 Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.141476 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.141762 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-log" containerID="cri-o://0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d" gracePeriod=30 Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.141839 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-metadata" containerID="cri-o://02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f" gracePeriod=30 Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.886528 4876 generic.go:334] "Generic (PLEG): container finished" podID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerID="0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d" exitCode=143 Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.886600 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1a13ae8-9c3c-43d6-af8e-cf13c3832166","Type":"ContainerDied","Data":"0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d"} Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.891966 4876 generic.go:334] "Generic (PLEG): container finished" podID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerID="8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf" exitCode=143 Dec 05 08:38:41 crc kubenswrapper[4876]: I1205 08:38:41.892013 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4e1018e-557a-47f3-bf3a-c2d62d83eece","Type":"ContainerDied","Data":"8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf"} Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.123096 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7156e3de784c88fc1a2fac9ecfd83d657be9d95e8cc28801d49c86222436e94" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.125579 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7156e3de784c88fc1a2fac9ecfd83d657be9d95e8cc28801d49c86222436e94" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.132135 4876 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7156e3de784c88fc1a2fac9ecfd83d657be9d95e8cc28801d49c86222436e94" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.132249 4876 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" containerName="nova-scheduler-scheduler" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.727014 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.742522 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-log-httpd\") pod \"4347cbb9-a742-4985-b4c8-a50c72d88a14\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.742588 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-sg-core-conf-yaml\") pod \"4347cbb9-a742-4985-b4c8-a50c72d88a14\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.742724 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-config-data\") pod \"4347cbb9-a742-4985-b4c8-a50c72d88a14\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.742753 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-run-httpd\") pod \"4347cbb9-a742-4985-b4c8-a50c72d88a14\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.742801 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-combined-ca-bundle\") pod \"4347cbb9-a742-4985-b4c8-a50c72d88a14\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.742934 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqtkn\" (UniqueName: \"kubernetes.io/projected/4347cbb9-a742-4985-b4c8-a50c72d88a14-kube-api-access-gqtkn\") pod \"4347cbb9-a742-4985-b4c8-a50c72d88a14\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.742969 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-scripts\") pod \"4347cbb9-a742-4985-b4c8-a50c72d88a14\" (UID: \"4347cbb9-a742-4985-b4c8-a50c72d88a14\") " Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.743123 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4347cbb9-a742-4985-b4c8-a50c72d88a14" (UID: "4347cbb9-a742-4985-b4c8-a50c72d88a14"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.743637 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4347cbb9-a742-4985-b4c8-a50c72d88a14" (UID: "4347cbb9-a742-4985-b4c8-a50c72d88a14"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.743688 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.750650 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-scripts" (OuterVolumeSpecName: "scripts") pod "4347cbb9-a742-4985-b4c8-a50c72d88a14" (UID: "4347cbb9-a742-4985-b4c8-a50c72d88a14"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.756227 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4347cbb9-a742-4985-b4c8-a50c72d88a14-kube-api-access-gqtkn" (OuterVolumeSpecName: "kube-api-access-gqtkn") pod "4347cbb9-a742-4985-b4c8-a50c72d88a14" (UID: "4347cbb9-a742-4985-b4c8-a50c72d88a14"). InnerVolumeSpecName "kube-api-access-gqtkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.783693 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4347cbb9-a742-4985-b4c8-a50c72d88a14" (UID: "4347cbb9-a742-4985-b4c8-a50c72d88a14"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.819459 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4347cbb9-a742-4985-b4c8-a50c72d88a14" (UID: "4347cbb9-a742-4985-b4c8-a50c72d88a14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.841592 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-config-data" (OuterVolumeSpecName: "config-data") pod "4347cbb9-a742-4985-b4c8-a50c72d88a14" (UID: "4347cbb9-a742-4985-b4c8-a50c72d88a14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.845789 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.845834 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.845843 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4347cbb9-a742-4985-b4c8-a50c72d88a14-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.845853 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.845862 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqtkn\" (UniqueName: \"kubernetes.io/projected/4347cbb9-a742-4985-b4c8-a50c72d88a14-kube-api-access-gqtkn\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.845873 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4347cbb9-a742-4985-b4c8-a50c72d88a14-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.911967 4876 generic.go:334] "Generic (PLEG): container finished" podID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerID="104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9" exitCode=137 Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.912015 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerDied","Data":"104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9"} Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.912123 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.912138 4876 scope.go:117] "RemoveContainer" containerID="104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.912087 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4347cbb9-a742-4985-b4c8-a50c72d88a14","Type":"ContainerDied","Data":"22e3f0467be0882ca230a6aa9bc8bb7912fcd0ec96e095588b54f8094222ab90"} Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.939974 4876 scope.go:117] "RemoveContainer" containerID="bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.951298 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.962075 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.984651 4876 scope.go:117] "RemoveContainer" containerID="9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.985366 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.985916 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerName="extract-content" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.985940 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerName="extract-content" Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.985960 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36f2a90b-0c85-478b-83e5-17ff10e01c94" containerName="nova-manage" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.985971 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="36f2a90b-0c85-478b-83e5-17ff10e01c94" containerName="nova-manage" Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.985986 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="ceilometer-central-agent" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.985994 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="ceilometer-central-agent" Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.986003 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="sg-core" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986010 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="sg-core" Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.986038 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerName="registry-server" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986046 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerName="registry-server" Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.986067 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="ceilometer-notification-agent" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986076 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="ceilometer-notification-agent" Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.986093 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="proxy-httpd" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986101 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="proxy-httpd" Dec 05 08:38:43 crc kubenswrapper[4876]: E1205 08:38:43.986117 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerName="extract-utilities" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986126 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerName="extract-utilities" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986364 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="ceilometer-notification-agent" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986380 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="sg-core" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986392 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="36f2a90b-0c85-478b-83e5-17ff10e01c94" containerName="nova-manage" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986403 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="ceilometer-central-agent" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986417 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="30bee729-e5e2-41a6-8aff-44f3efc4866f" containerName="registry-server" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.986440 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" containerName="proxy-httpd" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.988732 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.991078 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:38:43 crc kubenswrapper[4876]: I1205 08:38:43.991163 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.015536 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.024158 4876 scope.go:117] "RemoveContainer" containerID="2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.048655 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-config-data\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.048734 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-run-httpd\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.048780 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pff6d\" (UniqueName: \"kubernetes.io/projected/913fee04-2e39-4058-a9b5-8d064202f9ab-kube-api-access-pff6d\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.048852 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-scripts\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.048880 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.048938 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.049000 4876 scope.go:117] "RemoveContainer" containerID="104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.049128 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-log-httpd\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: E1205 08:38:44.049396 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9\": container with ID starting with 104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9 not found: ID does not exist" containerID="104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.049456 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9"} err="failed to get container status \"104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9\": rpc error: code = NotFound desc = could not find container \"104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9\": container with ID starting with 104dcb62be9ce40fde0c379dac1af49a7487b07905cb8fc70e6edb611c0eb4a9 not found: ID does not exist" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.049479 4876 scope.go:117] "RemoveContainer" containerID="bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb" Dec 05 08:38:44 crc kubenswrapper[4876]: E1205 08:38:44.049728 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb\": container with ID starting with bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb not found: ID does not exist" containerID="bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.049756 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb"} err="failed to get container status \"bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb\": rpc error: code = NotFound desc = could not find container \"bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb\": container with ID starting with bf5fa9cf909a23c6642c65c65334877be3866ed2cae3359927d60a60608668eb not found: ID does not exist" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.049771 4876 scope.go:117] "RemoveContainer" containerID="9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25" Dec 05 08:38:44 crc kubenswrapper[4876]: E1205 08:38:44.050011 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25\": container with ID starting with 9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25 not found: ID does not exist" containerID="9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.050071 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25"} err="failed to get container status \"9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25\": rpc error: code = NotFound desc = could not find container \"9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25\": container with ID starting with 9c1d524611e58bf73f04342176babd670bcac804ff574f56035a0f9953c9aa25 not found: ID does not exist" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.050086 4876 scope.go:117] "RemoveContainer" containerID="2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54" Dec 05 08:38:44 crc kubenswrapper[4876]: E1205 08:38:44.050256 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54\": container with ID starting with 2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54 not found: ID does not exist" containerID="2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.050280 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54"} err="failed to get container status \"2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54\": rpc error: code = NotFound desc = could not find container \"2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54\": container with ID starting with 2c470258cd996df90212d203c4c48d4133a91fd943b8e3b37866d93d27a3fc54 not found: ID does not exist" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.150379 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.150422 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.150471 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-log-httpd\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.150542 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-config-data\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.150609 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-run-httpd\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.150631 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pff6d\" (UniqueName: \"kubernetes.io/projected/913fee04-2e39-4058-a9b5-8d064202f9ab-kube-api-access-pff6d\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.150674 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-scripts\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.151797 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-log-httpd\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.152111 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-run-httpd\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.157252 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-scripts\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.158680 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.161815 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-config-data\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.163900 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.171573 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pff6d\" (UniqueName: \"kubernetes.io/projected/913fee04-2e39-4058-a9b5-8d064202f9ab-kube-api-access-pff6d\") pod \"ceilometer-0\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.325609 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.848610 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.853714 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.922073 4876 generic.go:334] "Generic (PLEG): container finished" podID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerID="d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8" exitCode=0 Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.922132 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4e1018e-557a-47f3-bf3a-c2d62d83eece","Type":"ContainerDied","Data":"d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8"} Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.922156 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4e1018e-557a-47f3-bf3a-c2d62d83eece","Type":"ContainerDied","Data":"6861dcc4778a4e95e625753095f20d51aa72b0a6c4c6d7c0ec0d1f05d3f205e3"} Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.922172 4876 scope.go:117] "RemoveContainer" containerID="d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.922263 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.926239 4876 generic.go:334] "Generic (PLEG): container finished" podID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerID="02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f" exitCode=0 Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.926361 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1a13ae8-9c3c-43d6-af8e-cf13c3832166","Type":"ContainerDied","Data":"02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f"} Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.926400 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c1a13ae8-9c3c-43d6-af8e-cf13c3832166","Type":"ContainerDied","Data":"71786f68c13dcfd6bea4de96d9c2a34b2145733180338aa1ce8abf0dc4e3823b"} Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.926486 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.948162 4876 scope.go:117] "RemoveContainer" containerID="8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.962507 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.978779 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-config-data\") pod \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.978861 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-combined-ca-bundle\") pod \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.978918 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-logs\") pod \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.978937 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-config-data\") pod \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.978963 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhpq4\" (UniqueName: \"kubernetes.io/projected/a4e1018e-557a-47f3-bf3a-c2d62d83eece-kube-api-access-dhpq4\") pod \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.979024 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qrsn\" (UniqueName: \"kubernetes.io/projected/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-kube-api-access-4qrsn\") pod \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.979066 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4e1018e-557a-47f3-bf3a-c2d62d83eece-logs\") pod \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.979348 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-nova-metadata-tls-certs\") pod \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\" (UID: \"c1a13ae8-9c3c-43d6-af8e-cf13c3832166\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.979388 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-combined-ca-bundle\") pod \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\" (UID: \"a4e1018e-557a-47f3-bf3a-c2d62d83eece\") " Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.979953 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-logs" (OuterVolumeSpecName: "logs") pod "c1a13ae8-9c3c-43d6-af8e-cf13c3832166" (UID: "c1a13ae8-9c3c-43d6-af8e-cf13c3832166"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.980036 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4e1018e-557a-47f3-bf3a-c2d62d83eece-logs" (OuterVolumeSpecName: "logs") pod "a4e1018e-557a-47f3-bf3a-c2d62d83eece" (UID: "a4e1018e-557a-47f3-bf3a-c2d62d83eece"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.984509 4876 scope.go:117] "RemoveContainer" containerID="d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.984552 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e1018e-557a-47f3-bf3a-c2d62d83eece-kube-api-access-dhpq4" (OuterVolumeSpecName: "kube-api-access-dhpq4") pod "a4e1018e-557a-47f3-bf3a-c2d62d83eece" (UID: "a4e1018e-557a-47f3-bf3a-c2d62d83eece"). InnerVolumeSpecName "kube-api-access-dhpq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.984598 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-kube-api-access-4qrsn" (OuterVolumeSpecName: "kube-api-access-4qrsn") pod "c1a13ae8-9c3c-43d6-af8e-cf13c3832166" (UID: "c1a13ae8-9c3c-43d6-af8e-cf13c3832166"). InnerVolumeSpecName "kube-api-access-4qrsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:44 crc kubenswrapper[4876]: E1205 08:38:44.986063 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8\": container with ID starting with d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8 not found: ID does not exist" containerID="d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.986114 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8"} err="failed to get container status \"d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8\": rpc error: code = NotFound desc = could not find container \"d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8\": container with ID starting with d722d377eda72c1914a8c2370dcbb94dc0a67f6fc7ae6a1876a763c4e5d961d8 not found: ID does not exist" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.986151 4876 scope.go:117] "RemoveContainer" containerID="8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf" Dec 05 08:38:44 crc kubenswrapper[4876]: E1205 08:38:44.987299 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf\": container with ID starting with 8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf not found: ID does not exist" containerID="8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.987332 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf"} err="failed to get container status \"8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf\": rpc error: code = NotFound desc = could not find container \"8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf\": container with ID starting with 8910c8f042171ad1b5440b8ec6efbf50f2009ae3e141f3c066d84b77cb5703cf not found: ID does not exist" Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.987348 4876 scope.go:117] "RemoveContainer" containerID="02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f" Dec 05 08:38:44 crc kubenswrapper[4876]: W1205 08:38:44.988549 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod913fee04_2e39_4058_a9b5_8d064202f9ab.slice/crio-6fe7f445532ff498345122e73ce27e7910df243f881d1bb42b468286801376d1 WatchSource:0}: Error finding container 6fe7f445532ff498345122e73ce27e7910df243f881d1bb42b468286801376d1: Status 404 returned error can't find the container with id 6fe7f445532ff498345122e73ce27e7910df243f881d1bb42b468286801376d1 Dec 05 08:38:44 crc kubenswrapper[4876]: I1205 08:38:44.991100 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.009191 4876 scope.go:117] "RemoveContainer" containerID="0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.009937 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1a13ae8-9c3c-43d6-af8e-cf13c3832166" (UID: "c1a13ae8-9c3c-43d6-af8e-cf13c3832166"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.014416 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-config-data" (OuterVolumeSpecName: "config-data") pod "a4e1018e-557a-47f3-bf3a-c2d62d83eece" (UID: "a4e1018e-557a-47f3-bf3a-c2d62d83eece"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.014448 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4e1018e-557a-47f3-bf3a-c2d62d83eece" (UID: "a4e1018e-557a-47f3-bf3a-c2d62d83eece"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.016644 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-config-data" (OuterVolumeSpecName: "config-data") pod "c1a13ae8-9c3c-43d6-af8e-cf13c3832166" (UID: "c1a13ae8-9c3c-43d6-af8e-cf13c3832166"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.029188 4876 scope.go:117] "RemoveContainer" containerID="02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f" Dec 05 08:38:45 crc kubenswrapper[4876]: E1205 08:38:45.029735 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f\": container with ID starting with 02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f not found: ID does not exist" containerID="02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.029781 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f"} err="failed to get container status \"02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f\": rpc error: code = NotFound desc = could not find container \"02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f\": container with ID starting with 02d48ec28afa900fd40734cea9e54fa31951d16bedf5981841a9af97e110b13f not found: ID does not exist" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.029811 4876 scope.go:117] "RemoveContainer" containerID="0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d" Dec 05 08:38:45 crc kubenswrapper[4876]: E1205 08:38:45.030099 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d\": container with ID starting with 0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d not found: ID does not exist" containerID="0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.030128 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d"} err="failed to get container status \"0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d\": rpc error: code = NotFound desc = could not find container \"0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d\": container with ID starting with 0d2fc9e3b7c0c414a02f1557147973bcd62daccb0d9d02e538d4aa3f8806920d not found: ID does not exist" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.042141 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c1a13ae8-9c3c-43d6-af8e-cf13c3832166" (UID: "c1a13ae8-9c3c-43d6-af8e-cf13c3832166"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081688 4876 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081727 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081738 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4e1018e-557a-47f3-bf3a-c2d62d83eece-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081751 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081762 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081775 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081785 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhpq4\" (UniqueName: \"kubernetes.io/projected/a4e1018e-557a-47f3-bf3a-c2d62d83eece-kube-api-access-dhpq4\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081797 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qrsn\" (UniqueName: \"kubernetes.io/projected/c1a13ae8-9c3c-43d6-af8e-cf13c3832166-kube-api-access-4qrsn\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.081810 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4e1018e-557a-47f3-bf3a-c2d62d83eece-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.260623 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.276638 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.320261 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.338188 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.361357 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:45 crc kubenswrapper[4876]: E1205 08:38:45.362225 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-metadata" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.362248 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-metadata" Dec 05 08:38:45 crc kubenswrapper[4876]: E1205 08:38:45.362264 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-api" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.362270 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-api" Dec 05 08:38:45 crc kubenswrapper[4876]: E1205 08:38:45.362283 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-log" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.362290 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-log" Dec 05 08:38:45 crc kubenswrapper[4876]: E1205 08:38:45.362302 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-log" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.362308 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-log" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.362495 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-metadata" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.362507 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" containerName="nova-metadata-log" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.362519 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-api" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.362534 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" containerName="nova-api-log" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.363584 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.365379 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.384449 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.395767 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.398208 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.400405 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.401332 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.405545 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.487867 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80bf9350-e2e3-429c-b7ad-b279ad895c5b-logs\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.487984 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.488018 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwzmf\" (UniqueName: \"kubernetes.io/projected/80bf9350-e2e3-429c-b7ad-b279ad895c5b-kube-api-access-hwzmf\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.488077 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-config-data\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589229 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae3be31-2c00-4486-9c64-10c7fcb2c787-logs\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589577 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gz62\" (UniqueName: \"kubernetes.io/projected/7ae3be31-2c00-4486-9c64-10c7fcb2c787-kube-api-access-6gz62\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589614 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589686 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80bf9350-e2e3-429c-b7ad-b279ad895c5b-logs\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589731 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-config-data\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589770 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589795 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwzmf\" (UniqueName: \"kubernetes.io/projected/80bf9350-e2e3-429c-b7ad-b279ad895c5b-kube-api-access-hwzmf\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589855 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.589888 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-config-data\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.591073 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80bf9350-e2e3-429c-b7ad-b279ad895c5b-logs\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.593703 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.593762 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-config-data\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.607594 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwzmf\" (UniqueName: \"kubernetes.io/projected/80bf9350-e2e3-429c-b7ad-b279ad895c5b-kube-api-access-hwzmf\") pod \"nova-api-0\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.716532 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.716713 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae3be31-2c00-4486-9c64-10c7fcb2c787-logs\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.716745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gz62\" (UniqueName: \"kubernetes.io/projected/7ae3be31-2c00-4486-9c64-10c7fcb2c787-kube-api-access-6gz62\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.716776 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.716879 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-config-data\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.717236 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ae3be31-2c00-4486-9c64-10c7fcb2c787-logs\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.721138 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.722075 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.724139 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ae3be31-2c00-4486-9c64-10c7fcb2c787-config-data\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.736945 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gz62\" (UniqueName: \"kubernetes.io/projected/7ae3be31-2c00-4486-9c64-10c7fcb2c787-kube-api-access-6gz62\") pod \"nova-metadata-0\" (UID: \"7ae3be31-2c00-4486-9c64-10c7fcb2c787\") " pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.835841 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4347cbb9-a742-4985-b4c8-a50c72d88a14" path="/var/lib/kubelet/pods/4347cbb9-a742-4985-b4c8-a50c72d88a14/volumes" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.836740 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4e1018e-557a-47f3-bf3a-c2d62d83eece" path="/var/lib/kubelet/pods/a4e1018e-557a-47f3-bf3a-c2d62d83eece/volumes" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.838104 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1a13ae8-9c3c-43d6-af8e-cf13c3832166" path="/var/lib/kubelet/pods/c1a13ae8-9c3c-43d6-af8e-cf13c3832166/volumes" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.841654 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.872779 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 08:38:45 crc kubenswrapper[4876]: I1205 08:38:45.956098 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerStarted","Data":"6fe7f445532ff498345122e73ce27e7910df243f881d1bb42b468286801376d1"} Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.351776 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.394550 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 08:38:46 crc kubenswrapper[4876]: W1205 08:38:46.409371 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ae3be31_2c00_4486_9c64_10c7fcb2c787.slice/crio-71641d2ce759e67958ff358a7d0e842595d6507bbfb64c47288bc3b65fa5f573 WatchSource:0}: Error finding container 71641d2ce759e67958ff358a7d0e842595d6507bbfb64c47288bc3b65fa5f573: Status 404 returned error can't find the container with id 71641d2ce759e67958ff358a7d0e842595d6507bbfb64c47288bc3b65fa5f573 Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.975359 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80bf9350-e2e3-429c-b7ad-b279ad895c5b","Type":"ContainerStarted","Data":"74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854"} Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.975860 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80bf9350-e2e3-429c-b7ad-b279ad895c5b","Type":"ContainerStarted","Data":"079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6"} Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.975876 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80bf9350-e2e3-429c-b7ad-b279ad895c5b","Type":"ContainerStarted","Data":"57aa09509012284a0e571e439a418f6cad519c88c4265ee8dcda96f5f58477dd"} Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.983918 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerStarted","Data":"6198cca04be028ebc2e2eb7f9a39a01f1b63c51d91cfc763a1cf412156b60ad0"} Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.983961 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerStarted","Data":"ee8dec0c913f9fafb0012c181b0eba2cf165b5031b68ace0a65759bba565cb56"} Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.994332 4876 generic.go:334] "Generic (PLEG): container finished" podID="e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" containerID="b7156e3de784c88fc1a2fac9ecfd83d657be9d95e8cc28801d49c86222436e94" exitCode=0 Dec 05 08:38:46 crc kubenswrapper[4876]: I1205 08:38:46.994407 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad","Type":"ContainerDied","Data":"b7156e3de784c88fc1a2fac9ecfd83d657be9d95e8cc28801d49c86222436e94"} Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:46.997553 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.997540755 podStartE2EDuration="1.997540755s" podCreationTimestamp="2025-12-05 08:38:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:46.996438683 +0000 UTC m=+1391.485103295" watchObservedRunningTime="2025-12-05 08:38:46.997540755 +0000 UTC m=+1391.486205377" Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.008701 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7ae3be31-2c00-4486-9c64-10c7fcb2c787","Type":"ContainerStarted","Data":"ff2e773321995dd927a4ec341eb94ebb9c73569e4ed39968f585c7edf25c37cb"} Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.008753 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7ae3be31-2c00-4486-9c64-10c7fcb2c787","Type":"ContainerStarted","Data":"db6a2fdcd5888362e65c27b8e50191b15f1b9d221066766d8efee6243da20a35"} Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.008766 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7ae3be31-2c00-4486-9c64-10c7fcb2c787","Type":"ContainerStarted","Data":"71641d2ce759e67958ff358a7d0e842595d6507bbfb64c47288bc3b65fa5f573"} Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.035107 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.03508153 podStartE2EDuration="2.03508153s" podCreationTimestamp="2025-12-05 08:38:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:47.027178713 +0000 UTC m=+1391.515843355" watchObservedRunningTime="2025-12-05 08:38:47.03508153 +0000 UTC m=+1391.523746152" Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.187080 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.365582 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqjw8\" (UniqueName: \"kubernetes.io/projected/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-kube-api-access-sqjw8\") pod \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.365652 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-config-data\") pod \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.365776 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-combined-ca-bundle\") pod \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\" (UID: \"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad\") " Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.388048 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-kube-api-access-sqjw8" (OuterVolumeSpecName: "kube-api-access-sqjw8") pod "e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" (UID: "e7f91add-96d8-4d36-a28c-e7a7c7cdfbad"). InnerVolumeSpecName "kube-api-access-sqjw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.419048 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-config-data" (OuterVolumeSpecName: "config-data") pod "e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" (UID: "e7f91add-96d8-4d36-a28c-e7a7c7cdfbad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.428038 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" (UID: "e7f91add-96d8-4d36-a28c-e7a7c7cdfbad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.468232 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.468273 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqjw8\" (UniqueName: \"kubernetes.io/projected/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-kube-api-access-sqjw8\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:47 crc kubenswrapper[4876]: I1205 08:38:47.468286 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.023651 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e7f91add-96d8-4d36-a28c-e7a7c7cdfbad","Type":"ContainerDied","Data":"f3b4a2e05f6ea3ab1959021c380d1ed3fe350a0a46f86d78be943c2d210634fb"} Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.023681 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.023740 4876 scope.go:117] "RemoveContainer" containerID="b7156e3de784c88fc1a2fac9ecfd83d657be9d95e8cc28801d49c86222436e94" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.031366 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerStarted","Data":"3a133e6e10243e2817f035a7e4973a15d1d85e9886f804885e1caaab23abea6f"} Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.050676 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.062474 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.074272 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:48 crc kubenswrapper[4876]: E1205 08:38:48.074857 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" containerName="nova-scheduler-scheduler" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.075430 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" containerName="nova-scheduler-scheduler" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.075778 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" containerName="nova-scheduler-scheduler" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.076721 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.078667 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.094242 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.182174 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304bb281-a6d7-460a-a81e-3cecbb60e20f-config-data\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.182254 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304bb281-a6d7-460a-a81e-3cecbb60e20f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.182359 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ssxv\" (UniqueName: \"kubernetes.io/projected/304bb281-a6d7-460a-a81e-3cecbb60e20f-kube-api-access-8ssxv\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.284329 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304bb281-a6d7-460a-a81e-3cecbb60e20f-config-data\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.284796 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304bb281-a6d7-460a-a81e-3cecbb60e20f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.284957 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ssxv\" (UniqueName: \"kubernetes.io/projected/304bb281-a6d7-460a-a81e-3cecbb60e20f-kube-api-access-8ssxv\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.290497 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304bb281-a6d7-460a-a81e-3cecbb60e20f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.290577 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304bb281-a6d7-460a-a81e-3cecbb60e20f-config-data\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.306072 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ssxv\" (UniqueName: \"kubernetes.io/projected/304bb281-a6d7-460a-a81e-3cecbb60e20f-kube-api-access-8ssxv\") pod \"nova-scheduler-0\" (UID: \"304bb281-a6d7-460a-a81e-3cecbb60e20f\") " pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.392669 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 08:38:48 crc kubenswrapper[4876]: I1205 08:38:48.816031 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 08:38:48 crc kubenswrapper[4876]: W1205 08:38:48.825617 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod304bb281_a6d7_460a_a81e_3cecbb60e20f.slice/crio-8efa5b5de1f7802fbdcb63d630c1e3afd4f128120dbf764f02fcb1e7e26ef657 WatchSource:0}: Error finding container 8efa5b5de1f7802fbdcb63d630c1e3afd4f128120dbf764f02fcb1e7e26ef657: Status 404 returned error can't find the container with id 8efa5b5de1f7802fbdcb63d630c1e3afd4f128120dbf764f02fcb1e7e26ef657 Dec 05 08:38:49 crc kubenswrapper[4876]: I1205 08:38:49.042735 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerStarted","Data":"b1ad66f249bee461d27fec10eebc20e6160b0a216730d97bfdd0b812c22d0e60"} Dec 05 08:38:49 crc kubenswrapper[4876]: I1205 08:38:49.043164 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:38:49 crc kubenswrapper[4876]: I1205 08:38:49.044258 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"304bb281-a6d7-460a-a81e-3cecbb60e20f","Type":"ContainerStarted","Data":"329287409ec9a3463f5c8b8bffe3027a595d15cf0377e8222b82485cb0a0b2c9"} Dec 05 08:38:49 crc kubenswrapper[4876]: I1205 08:38:49.044297 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"304bb281-a6d7-460a-a81e-3cecbb60e20f","Type":"ContainerStarted","Data":"8efa5b5de1f7802fbdcb63d630c1e3afd4f128120dbf764f02fcb1e7e26ef657"} Dec 05 08:38:49 crc kubenswrapper[4876]: I1205 08:38:49.063123 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.383485025 podStartE2EDuration="6.063103958s" podCreationTimestamp="2025-12-05 08:38:43 +0000 UTC" firstStartedPulling="2025-12-05 08:38:44.99082107 +0000 UTC m=+1389.479485682" lastFinishedPulling="2025-12-05 08:38:48.670439993 +0000 UTC m=+1393.159104615" observedRunningTime="2025-12-05 08:38:49.062770751 +0000 UTC m=+1393.551435373" watchObservedRunningTime="2025-12-05 08:38:49.063103958 +0000 UTC m=+1393.551768580" Dec 05 08:38:49 crc kubenswrapper[4876]: I1205 08:38:49.836541 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f91add-96d8-4d36-a28c-e7a7c7cdfbad" path="/var/lib/kubelet/pods/e7f91add-96d8-4d36-a28c-e7a7c7cdfbad/volumes" Dec 05 08:38:50 crc kubenswrapper[4876]: I1205 08:38:50.873807 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:38:50 crc kubenswrapper[4876]: I1205 08:38:50.874368 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 08:38:53 crc kubenswrapper[4876]: I1205 08:38:53.393557 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 08:38:55 crc kubenswrapper[4876]: I1205 08:38:55.842689 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:38:55 crc kubenswrapper[4876]: I1205 08:38:55.842980 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:38:55 crc kubenswrapper[4876]: I1205 08:38:55.874135 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:38:55 crc kubenswrapper[4876]: I1205 08:38:55.874521 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 08:38:56 crc kubenswrapper[4876]: I1205 08:38:56.940198 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:56 crc kubenswrapper[4876]: I1205 08:38:56.940257 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:56 crc kubenswrapper[4876]: I1205 08:38:56.940318 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7ae3be31-2c00-4486-9c64-10c7fcb2c787" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:56 crc kubenswrapper[4876]: I1205 08:38:56.942084 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="7ae3be31-2c00-4486-9c64-10c7fcb2c787" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 08:38:58 crc kubenswrapper[4876]: I1205 08:38:58.393096 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 08:38:58 crc kubenswrapper[4876]: I1205 08:38:58.425635 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 08:38:58 crc kubenswrapper[4876]: I1205 08:38:58.446612 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=10.446592728 podStartE2EDuration="10.446592728s" podCreationTimestamp="2025-12-05 08:38:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:38:49.085702196 +0000 UTC m=+1393.574366818" watchObservedRunningTime="2025-12-05 08:38:58.446592728 +0000 UTC m=+1402.935257350" Dec 05 08:38:59 crc kubenswrapper[4876]: I1205 08:38:59.191121 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 08:39:05 crc kubenswrapper[4876]: I1205 08:39:05.847508 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 08:39:05 crc kubenswrapper[4876]: I1205 08:39:05.848786 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 08:39:05 crc kubenswrapper[4876]: I1205 08:39:05.850460 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 08:39:05 crc kubenswrapper[4876]: I1205 08:39:05.854464 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 08:39:05 crc kubenswrapper[4876]: I1205 08:39:05.878366 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 08:39:05 crc kubenswrapper[4876]: I1205 08:39:05.881519 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 08:39:05 crc kubenswrapper[4876]: I1205 08:39:05.881770 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.222928 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.229120 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.233997 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.448007 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-tmh27"] Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.450134 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.480784 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-tmh27"] Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.543850 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-config\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.543935 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.544018 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.544061 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.544141 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24lt7\" (UniqueName: \"kubernetes.io/projected/2db748c5-cab9-4494-8ba0-ff71f7b8b252-kube-api-access-24lt7\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.544200 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.645634 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.645700 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.645791 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24lt7\" (UniqueName: \"kubernetes.io/projected/2db748c5-cab9-4494-8ba0-ff71f7b8b252-kube-api-access-24lt7\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.645841 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.646067 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-config\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.646123 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.647152 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.647591 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.648104 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-config\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.648569 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.650915 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.667222 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24lt7\" (UniqueName: \"kubernetes.io/projected/2db748c5-cab9-4494-8ba0-ff71f7b8b252-kube-api-access-24lt7\") pod \"dnsmasq-dns-6d99f6bc7f-tmh27\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:06 crc kubenswrapper[4876]: I1205 08:39:06.778855 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:07 crc kubenswrapper[4876]: I1205 08:39:07.259626 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-tmh27"] Dec 05 08:39:07 crc kubenswrapper[4876]: W1205 08:39:07.264050 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2db748c5_cab9_4494_8ba0_ff71f7b8b252.slice/crio-a282cc6589c82d98e6a415be0894ff9b3401e543bbd8f68c3dbc7e908bdb55bc WatchSource:0}: Error finding container a282cc6589c82d98e6a415be0894ff9b3401e543bbd8f68c3dbc7e908bdb55bc: Status 404 returned error can't find the container with id a282cc6589c82d98e6a415be0894ff9b3401e543bbd8f68c3dbc7e908bdb55bc Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.244046 4876 generic.go:334] "Generic (PLEG): container finished" podID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" containerID="0c6280ae9e8c0648a1c1b8d7d9bb7e9444a1282df9ea731bfc4f64968c18b5ec" exitCode=0 Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.244121 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" event={"ID":"2db748c5-cab9-4494-8ba0-ff71f7b8b252","Type":"ContainerDied","Data":"0c6280ae9e8c0648a1c1b8d7d9bb7e9444a1282df9ea731bfc4f64968c18b5ec"} Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.246845 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" event={"ID":"2db748c5-cab9-4494-8ba0-ff71f7b8b252","Type":"ContainerStarted","Data":"a282cc6589c82d98e6a415be0894ff9b3401e543bbd8f68c3dbc7e908bdb55bc"} Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.342541 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.342885 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="ceilometer-central-agent" containerID="cri-o://ee8dec0c913f9fafb0012c181b0eba2cf165b5031b68ace0a65759bba565cb56" gracePeriod=30 Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.343062 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="proxy-httpd" containerID="cri-o://b1ad66f249bee461d27fec10eebc20e6160b0a216730d97bfdd0b812c22d0e60" gracePeriod=30 Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.343115 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="sg-core" containerID="cri-o://3a133e6e10243e2817f035a7e4973a15d1d85e9886f804885e1caaab23abea6f" gracePeriod=30 Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.343153 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="ceilometer-notification-agent" containerID="cri-o://6198cca04be028ebc2e2eb7f9a39a01f1b63c51d91cfc763a1cf412156b60ad0" gracePeriod=30 Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.447786 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.204:3000/\": read tcp 10.217.0.2:32770->10.217.0.204:3000: read: connection reset by peer" Dec 05 08:39:08 crc kubenswrapper[4876]: I1205 08:39:08.932111 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.259271 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" event={"ID":"2db748c5-cab9-4494-8ba0-ff71f7b8b252","Type":"ContainerStarted","Data":"6c3ff6b383912f24151094bdbc91f3b7716fc94c398919c79a9b60cf6e5a50ea"} Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.259624 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.262321 4876 generic.go:334] "Generic (PLEG): container finished" podID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerID="b1ad66f249bee461d27fec10eebc20e6160b0a216730d97bfdd0b812c22d0e60" exitCode=0 Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.262362 4876 generic.go:334] "Generic (PLEG): container finished" podID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerID="3a133e6e10243e2817f035a7e4973a15d1d85e9886f804885e1caaab23abea6f" exitCode=2 Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.262377 4876 generic.go:334] "Generic (PLEG): container finished" podID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerID="ee8dec0c913f9fafb0012c181b0eba2cf165b5031b68ace0a65759bba565cb56" exitCode=0 Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.262420 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerDied","Data":"b1ad66f249bee461d27fec10eebc20e6160b0a216730d97bfdd0b812c22d0e60"} Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.262474 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerDied","Data":"3a133e6e10243e2817f035a7e4973a15d1d85e9886f804885e1caaab23abea6f"} Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.262487 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerDied","Data":"ee8dec0c913f9fafb0012c181b0eba2cf165b5031b68ace0a65759bba565cb56"} Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.262571 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-log" containerID="cri-o://079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6" gracePeriod=30 Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.262622 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-api" containerID="cri-o://74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854" gracePeriod=30 Dec 05 08:39:09 crc kubenswrapper[4876]: I1205 08:39:09.282003 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" podStartSLOduration=3.281976467 podStartE2EDuration="3.281976467s" podCreationTimestamp="2025-12-05 08:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:39:09.276594264 +0000 UTC m=+1413.765258896" watchObservedRunningTime="2025-12-05 08:39:09.281976467 +0000 UTC m=+1413.770641099" Dec 05 08:39:10 crc kubenswrapper[4876]: I1205 08:39:10.273611 4876 generic.go:334] "Generic (PLEG): container finished" podID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerID="079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6" exitCode=143 Dec 05 08:39:10 crc kubenswrapper[4876]: I1205 08:39:10.273696 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80bf9350-e2e3-429c-b7ad-b279ad895c5b","Type":"ContainerDied","Data":"079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6"} Dec 05 08:39:12 crc kubenswrapper[4876]: I1205 08:39:12.859594 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:39:12 crc kubenswrapper[4876]: I1205 08:39:12.991317 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwzmf\" (UniqueName: \"kubernetes.io/projected/80bf9350-e2e3-429c-b7ad-b279ad895c5b-kube-api-access-hwzmf\") pod \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " Dec 05 08:39:12 crc kubenswrapper[4876]: I1205 08:39:12.991371 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-config-data\") pod \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " Dec 05 08:39:12 crc kubenswrapper[4876]: I1205 08:39:12.991398 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80bf9350-e2e3-429c-b7ad-b279ad895c5b-logs\") pod \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " Dec 05 08:39:12 crc kubenswrapper[4876]: I1205 08:39:12.991612 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-combined-ca-bundle\") pod \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\" (UID: \"80bf9350-e2e3-429c-b7ad-b279ad895c5b\") " Dec 05 08:39:12 crc kubenswrapper[4876]: I1205 08:39:12.992107 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80bf9350-e2e3-429c-b7ad-b279ad895c5b-logs" (OuterVolumeSpecName: "logs") pod "80bf9350-e2e3-429c-b7ad-b279ad895c5b" (UID: "80bf9350-e2e3-429c-b7ad-b279ad895c5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:39:12 crc kubenswrapper[4876]: I1205 08:39:12.994055 4876 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80bf9350-e2e3-429c-b7ad-b279ad895c5b-logs\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:12 crc kubenswrapper[4876]: I1205 08:39:12.997505 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80bf9350-e2e3-429c-b7ad-b279ad895c5b-kube-api-access-hwzmf" (OuterVolumeSpecName: "kube-api-access-hwzmf") pod "80bf9350-e2e3-429c-b7ad-b279ad895c5b" (UID: "80bf9350-e2e3-429c-b7ad-b279ad895c5b"). InnerVolumeSpecName "kube-api-access-hwzmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.021074 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-config-data" (OuterVolumeSpecName: "config-data") pod "80bf9350-e2e3-429c-b7ad-b279ad895c5b" (UID: "80bf9350-e2e3-429c-b7ad-b279ad895c5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.034092 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80bf9350-e2e3-429c-b7ad-b279ad895c5b" (UID: "80bf9350-e2e3-429c-b7ad-b279ad895c5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.096778 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwzmf\" (UniqueName: \"kubernetes.io/projected/80bf9350-e2e3-429c-b7ad-b279ad895c5b-kube-api-access-hwzmf\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.096832 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.096845 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80bf9350-e2e3-429c-b7ad-b279ad895c5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.314236 4876 generic.go:334] "Generic (PLEG): container finished" podID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerID="6198cca04be028ebc2e2eb7f9a39a01f1b63c51d91cfc763a1cf412156b60ad0" exitCode=0 Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.314314 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerDied","Data":"6198cca04be028ebc2e2eb7f9a39a01f1b63c51d91cfc763a1cf412156b60ad0"} Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.316702 4876 generic.go:334] "Generic (PLEG): container finished" podID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerID="74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854" exitCode=0 Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.316740 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80bf9350-e2e3-429c-b7ad-b279ad895c5b","Type":"ContainerDied","Data":"74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854"} Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.316764 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"80bf9350-e2e3-429c-b7ad-b279ad895c5b","Type":"ContainerDied","Data":"57aa09509012284a0e571e439a418f6cad519c88c4265ee8dcda96f5f58477dd"} Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.316795 4876 scope.go:117] "RemoveContainer" containerID="74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.317094 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.337000 4876 scope.go:117] "RemoveContainer" containerID="079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.359617 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.373116 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.379467 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.381498 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 08:39:13 crc kubenswrapper[4876]: E1205 08:39:13.381817 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-log" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.381827 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-log" Dec 05 08:39:13 crc kubenswrapper[4876]: E1205 08:39:13.381844 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="ceilometer-central-agent" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.381851 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="ceilometer-central-agent" Dec 05 08:39:13 crc kubenswrapper[4876]: E1205 08:39:13.381874 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="ceilometer-notification-agent" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.381880 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="ceilometer-notification-agent" Dec 05 08:39:13 crc kubenswrapper[4876]: E1205 08:39:13.381888 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="proxy-httpd" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.381894 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="proxy-httpd" Dec 05 08:39:13 crc kubenswrapper[4876]: E1205 08:39:13.381923 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-api" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.381930 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-api" Dec 05 08:39:13 crc kubenswrapper[4876]: E1205 08:39:13.381943 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="sg-core" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.381949 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="sg-core" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.382115 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-log" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.382134 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="proxy-httpd" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.382144 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="sg-core" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.382153 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" containerName="nova-api-api" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.382164 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="ceilometer-central-agent" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.382174 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" containerName="ceilometer-notification-agent" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.383189 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.385607 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.385819 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.385960 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.390162 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.391281 4876 scope.go:117] "RemoveContainer" containerID="74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854" Dec 05 08:39:13 crc kubenswrapper[4876]: E1205 08:39:13.397511 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854\": container with ID starting with 74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854 not found: ID does not exist" containerID="74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.397556 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854"} err="failed to get container status \"74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854\": rpc error: code = NotFound desc = could not find container \"74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854\": container with ID starting with 74565c75db9f1b04d2897a74e3a59378eb18f42c434d88ec601ef176f0a0a854 not found: ID does not exist" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.397585 4876 scope.go:117] "RemoveContainer" containerID="079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6" Dec 05 08:39:13 crc kubenswrapper[4876]: E1205 08:39:13.398385 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6\": container with ID starting with 079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6 not found: ID does not exist" containerID="079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.398441 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6"} err="failed to get container status \"079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6\": rpc error: code = NotFound desc = could not find container \"079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6\": container with ID starting with 079df8662af1f5b2c7e109225d3e1f166bd584112bc5438082dcfdf272179ec6 not found: ID does not exist" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511092 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-config-data\") pod \"913fee04-2e39-4058-a9b5-8d064202f9ab\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511137 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-scripts\") pod \"913fee04-2e39-4058-a9b5-8d064202f9ab\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511215 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-sg-core-conf-yaml\") pod \"913fee04-2e39-4058-a9b5-8d064202f9ab\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511244 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-run-httpd\") pod \"913fee04-2e39-4058-a9b5-8d064202f9ab\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511290 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pff6d\" (UniqueName: \"kubernetes.io/projected/913fee04-2e39-4058-a9b5-8d064202f9ab-kube-api-access-pff6d\") pod \"913fee04-2e39-4058-a9b5-8d064202f9ab\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511342 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-combined-ca-bundle\") pod \"913fee04-2e39-4058-a9b5-8d064202f9ab\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511368 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-log-httpd\") pod \"913fee04-2e39-4058-a9b5-8d064202f9ab\" (UID: \"913fee04-2e39-4058-a9b5-8d064202f9ab\") " Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511641 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8b58af0-0308-45aa-825c-0d252a0bff0d-logs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511713 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-config-data\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511765 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511817 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j24xr\" (UniqueName: \"kubernetes.io/projected/b8b58af0-0308-45aa-825c-0d252a0bff0d-kube-api-access-j24xr\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511836 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.511849 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-public-tls-certs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.512987 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "913fee04-2e39-4058-a9b5-8d064202f9ab" (UID: "913fee04-2e39-4058-a9b5-8d064202f9ab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.513569 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "913fee04-2e39-4058-a9b5-8d064202f9ab" (UID: "913fee04-2e39-4058-a9b5-8d064202f9ab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.517579 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-scripts" (OuterVolumeSpecName: "scripts") pod "913fee04-2e39-4058-a9b5-8d064202f9ab" (UID: "913fee04-2e39-4058-a9b5-8d064202f9ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.518449 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/913fee04-2e39-4058-a9b5-8d064202f9ab-kube-api-access-pff6d" (OuterVolumeSpecName: "kube-api-access-pff6d") pod "913fee04-2e39-4058-a9b5-8d064202f9ab" (UID: "913fee04-2e39-4058-a9b5-8d064202f9ab"). InnerVolumeSpecName "kube-api-access-pff6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.556735 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "913fee04-2e39-4058-a9b5-8d064202f9ab" (UID: "913fee04-2e39-4058-a9b5-8d064202f9ab"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.605924 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "913fee04-2e39-4058-a9b5-8d064202f9ab" (UID: "913fee04-2e39-4058-a9b5-8d064202f9ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614053 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-config-data\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614134 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614193 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j24xr\" (UniqueName: \"kubernetes.io/projected/b8b58af0-0308-45aa-825c-0d252a0bff0d-kube-api-access-j24xr\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614212 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614224 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-public-tls-certs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614263 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8b58af0-0308-45aa-825c-0d252a0bff0d-logs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614317 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pff6d\" (UniqueName: \"kubernetes.io/projected/913fee04-2e39-4058-a9b5-8d064202f9ab-kube-api-access-pff6d\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614328 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614336 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614344 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614355 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614363 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/913fee04-2e39-4058-a9b5-8d064202f9ab-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.614714 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8b58af0-0308-45aa-825c-0d252a0bff0d-logs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.619270 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-public-tls-certs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.619768 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.620269 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.631946 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8b58af0-0308-45aa-825c-0d252a0bff0d-config-data\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.634949 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j24xr\" (UniqueName: \"kubernetes.io/projected/b8b58af0-0308-45aa-825c-0d252a0bff0d-kube-api-access-j24xr\") pod \"nova-api-0\" (UID: \"b8b58af0-0308-45aa-825c-0d252a0bff0d\") " pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.641610 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-config-data" (OuterVolumeSpecName: "config-data") pod "913fee04-2e39-4058-a9b5-8d064202f9ab" (UID: "913fee04-2e39-4058-a9b5-8d064202f9ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.699970 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.716419 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/913fee04-2e39-4058-a9b5-8d064202f9ab-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:13 crc kubenswrapper[4876]: I1205 08:39:13.835636 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80bf9350-e2e3-429c-b7ad-b279ad895c5b" path="/var/lib/kubelet/pods/80bf9350-e2e3-429c-b7ad-b279ad895c5b/volumes" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.151063 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.329027 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"913fee04-2e39-4058-a9b5-8d064202f9ab","Type":"ContainerDied","Data":"6fe7f445532ff498345122e73ce27e7910df243f881d1bb42b468286801376d1"} Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.329348 4876 scope.go:117] "RemoveContainer" containerID="b1ad66f249bee461d27fec10eebc20e6160b0a216730d97bfdd0b812c22d0e60" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.329063 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.332344 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b8b58af0-0308-45aa-825c-0d252a0bff0d","Type":"ContainerStarted","Data":"c3a35c08de7c66904f32aaf783738e115d105ec0a8bff57a3100e5573d85495f"} Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.332415 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b8b58af0-0308-45aa-825c-0d252a0bff0d","Type":"ContainerStarted","Data":"ced4df6328b56eee8e3d14a15ebb7e438b56e21ef62ed1770c964f494be7d71f"} Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.356602 4876 scope.go:117] "RemoveContainer" containerID="3a133e6e10243e2817f035a7e4973a15d1d85e9886f804885e1caaab23abea6f" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.416250 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.440092 4876 scope.go:117] "RemoveContainer" containerID="6198cca04be028ebc2e2eb7f9a39a01f1b63c51d91cfc763a1cf412156b60ad0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.444891 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.454470 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.457751 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.460241 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.464452 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.465512 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.474724 4876 scope.go:117] "RemoveContainer" containerID="ee8dec0c913f9fafb0012c181b0eba2cf165b5031b68ace0a65759bba565cb56" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.539767 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-config-data\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.539809 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-run-httpd\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.539838 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.539930 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.539982 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25dbf\" (UniqueName: \"kubernetes.io/projected/239ff295-f4fb-4d94-b893-76c436b41015-kube-api-access-25dbf\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.540004 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-scripts\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.540044 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-log-httpd\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.643021 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25dbf\" (UniqueName: \"kubernetes.io/projected/239ff295-f4fb-4d94-b893-76c436b41015-kube-api-access-25dbf\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.643129 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-scripts\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.643204 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-log-httpd\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.643253 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-config-data\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.643279 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-run-httpd\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.643320 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.643402 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.644374 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-log-httpd\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.644495 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-run-httpd\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.649019 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.650589 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.650608 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-scripts\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.651015 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-config-data\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.665319 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25dbf\" (UniqueName: \"kubernetes.io/projected/239ff295-f4fb-4d94-b893-76c436b41015-kube-api-access-25dbf\") pod \"ceilometer-0\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " pod="openstack/ceilometer-0" Dec 05 08:39:14 crc kubenswrapper[4876]: I1205 08:39:14.918157 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:39:15 crc kubenswrapper[4876]: I1205 08:39:15.346433 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b8b58af0-0308-45aa-825c-0d252a0bff0d","Type":"ContainerStarted","Data":"0e5b4c46a8568c97a852ef129a95d473617f2ac43bf5ca8337ae70409735ff4d"} Dec 05 08:39:15 crc kubenswrapper[4876]: I1205 08:39:15.373695 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.373663428 podStartE2EDuration="2.373663428s" podCreationTimestamp="2025-12-05 08:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:39:15.365363667 +0000 UTC m=+1419.854028329" watchObservedRunningTime="2025-12-05 08:39:15.373663428 +0000 UTC m=+1419.862328060" Dec 05 08:39:15 crc kubenswrapper[4876]: W1205 08:39:15.422430 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod239ff295_f4fb_4d94_b893_76c436b41015.slice/crio-475ad07415119597e38d9c988d756565b5cb1792beb2885462b0951907bf4751 WatchSource:0}: Error finding container 475ad07415119597e38d9c988d756565b5cb1792beb2885462b0951907bf4751: Status 404 returned error can't find the container with id 475ad07415119597e38d9c988d756565b5cb1792beb2885462b0951907bf4751 Dec 05 08:39:15 crc kubenswrapper[4876]: I1205 08:39:15.425180 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:15 crc kubenswrapper[4876]: I1205 08:39:15.843804 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="913fee04-2e39-4058-a9b5-8d064202f9ab" path="/var/lib/kubelet/pods/913fee04-2e39-4058-a9b5-8d064202f9ab/volumes" Dec 05 08:39:16 crc kubenswrapper[4876]: I1205 08:39:16.371564 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerStarted","Data":"475ad07415119597e38d9c988d756565b5cb1792beb2885462b0951907bf4751"} Dec 05 08:39:16 crc kubenswrapper[4876]: I1205 08:39:16.780992 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:39:16 crc kubenswrapper[4876]: I1205 08:39:16.867455 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-pw6bc"] Dec 05 08:39:16 crc kubenswrapper[4876]: I1205 08:39:16.867728 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" podUID="14bc9681-6da3-4e8b-8a88-4940f7b359de" containerName="dnsmasq-dns" containerID="cri-o://5a342e34f983f5092466e9864f0f8c3cf28865539e000d238e11d3aaf9f6ab15" gracePeriod=10 Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.387775 4876 generic.go:334] "Generic (PLEG): container finished" podID="14bc9681-6da3-4e8b-8a88-4940f7b359de" containerID="5a342e34f983f5092466e9864f0f8c3cf28865539e000d238e11d3aaf9f6ab15" exitCode=0 Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.387842 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" event={"ID":"14bc9681-6da3-4e8b-8a88-4940f7b359de","Type":"ContainerDied","Data":"5a342e34f983f5092466e9864f0f8c3cf28865539e000d238e11d3aaf9f6ab15"} Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.388305 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" event={"ID":"14bc9681-6da3-4e8b-8a88-4940f7b359de","Type":"ContainerDied","Data":"5fbc934ad18e3a94648eb14a5f02e13db8c65f13ca458c53731e3665fd5f1ba1"} Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.388322 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fbc934ad18e3a94648eb14a5f02e13db8c65f13ca458c53731e3665fd5f1ba1" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.392365 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerStarted","Data":"11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd"} Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.392764 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerStarted","Data":"fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056"} Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.441775 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.507910 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-nb\") pod \"14bc9681-6da3-4e8b-8a88-4940f7b359de\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.507960 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxzmf\" (UniqueName: \"kubernetes.io/projected/14bc9681-6da3-4e8b-8a88-4940f7b359de-kube-api-access-rxzmf\") pod \"14bc9681-6da3-4e8b-8a88-4940f7b359de\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.507996 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-swift-storage-0\") pod \"14bc9681-6da3-4e8b-8a88-4940f7b359de\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.508053 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-config\") pod \"14bc9681-6da3-4e8b-8a88-4940f7b359de\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.508116 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-svc\") pod \"14bc9681-6da3-4e8b-8a88-4940f7b359de\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.508146 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-sb\") pod \"14bc9681-6da3-4e8b-8a88-4940f7b359de\" (UID: \"14bc9681-6da3-4e8b-8a88-4940f7b359de\") " Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.513020 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14bc9681-6da3-4e8b-8a88-4940f7b359de-kube-api-access-rxzmf" (OuterVolumeSpecName: "kube-api-access-rxzmf") pod "14bc9681-6da3-4e8b-8a88-4940f7b359de" (UID: "14bc9681-6da3-4e8b-8a88-4940f7b359de"). InnerVolumeSpecName "kube-api-access-rxzmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.561683 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "14bc9681-6da3-4e8b-8a88-4940f7b359de" (UID: "14bc9681-6da3-4e8b-8a88-4940f7b359de"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.595711 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14bc9681-6da3-4e8b-8a88-4940f7b359de" (UID: "14bc9681-6da3-4e8b-8a88-4940f7b359de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.606270 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14bc9681-6da3-4e8b-8a88-4940f7b359de" (UID: "14bc9681-6da3-4e8b-8a88-4940f7b359de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.610389 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxzmf\" (UniqueName: \"kubernetes.io/projected/14bc9681-6da3-4e8b-8a88-4940f7b359de-kube-api-access-rxzmf\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.610425 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.610436 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.610448 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.611799 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14bc9681-6da3-4e8b-8a88-4940f7b359de" (UID: "14bc9681-6da3-4e8b-8a88-4940f7b359de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.619744 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-config" (OuterVolumeSpecName: "config") pod "14bc9681-6da3-4e8b-8a88-4940f7b359de" (UID: "14bc9681-6da3-4e8b-8a88-4940f7b359de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.712884 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:17 crc kubenswrapper[4876]: I1205 08:39:17.712925 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14bc9681-6da3-4e8b-8a88-4940f7b359de-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:18 crc kubenswrapper[4876]: I1205 08:39:18.407860 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerStarted","Data":"230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5"} Dec 05 08:39:18 crc kubenswrapper[4876]: I1205 08:39:18.407906 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-pw6bc" Dec 05 08:39:18 crc kubenswrapper[4876]: I1205 08:39:18.438571 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-pw6bc"] Dec 05 08:39:18 crc kubenswrapper[4876]: I1205 08:39:18.448356 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-pw6bc"] Dec 05 08:39:19 crc kubenswrapper[4876]: I1205 08:39:19.838149 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14bc9681-6da3-4e8b-8a88-4940f7b359de" path="/var/lib/kubelet/pods/14bc9681-6da3-4e8b-8a88-4940f7b359de/volumes" Dec 05 08:39:20 crc kubenswrapper[4876]: I1205 08:39:20.434319 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerStarted","Data":"a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286"} Dec 05 08:39:20 crc kubenswrapper[4876]: I1205 08:39:20.434540 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:39:20 crc kubenswrapper[4876]: I1205 08:39:20.460493 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.435408981 podStartE2EDuration="6.46047657s" podCreationTimestamp="2025-12-05 08:39:14 +0000 UTC" firstStartedPulling="2025-12-05 08:39:15.42697425 +0000 UTC m=+1419.915638882" lastFinishedPulling="2025-12-05 08:39:19.452041849 +0000 UTC m=+1423.940706471" observedRunningTime="2025-12-05 08:39:20.454894181 +0000 UTC m=+1424.943558863" watchObservedRunningTime="2025-12-05 08:39:20.46047657 +0000 UTC m=+1424.949141192" Dec 05 08:39:23 crc kubenswrapper[4876]: I1205 08:39:23.701111 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:39:23 crc kubenswrapper[4876]: I1205 08:39:23.701737 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 08:39:24 crc kubenswrapper[4876]: I1205 08:39:24.754394 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b8b58af0-0308-45aa-825c-0d252a0bff0d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 08:39:24 crc kubenswrapper[4876]: I1205 08:39:24.758097 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b8b58af0-0308-45aa-825c-0d252a0bff0d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 08:39:33 crc kubenswrapper[4876]: I1205 08:39:33.710546 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 08:39:33 crc kubenswrapper[4876]: I1205 08:39:33.711427 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 08:39:33 crc kubenswrapper[4876]: I1205 08:39:33.711863 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 08:39:33 crc kubenswrapper[4876]: I1205 08:39:33.711926 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 08:39:33 crc kubenswrapper[4876]: I1205 08:39:33.720539 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 08:39:33 crc kubenswrapper[4876]: I1205 08:39:33.723554 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 08:39:38 crc kubenswrapper[4876]: I1205 08:39:38.214050 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:39:38 crc kubenswrapper[4876]: I1205 08:39:38.214661 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:39:44 crc kubenswrapper[4876]: I1205 08:39:44.926876 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 08:39:48 crc kubenswrapper[4876]: I1205 08:39:48.383620 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:39:48 crc kubenswrapper[4876]: I1205 08:39:48.384184 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="b29eddc0-b7dd-4390-ab85-0e8eb61422e2" containerName="kube-state-metrics" containerID="cri-o://4a541af3fa0c88410e322db126e7852014165b17d79a5029bf58c38739650f88" gracePeriod=30 Dec 05 08:39:48 crc kubenswrapper[4876]: I1205 08:39:48.698594 4876 generic.go:334] "Generic (PLEG): container finished" podID="b29eddc0-b7dd-4390-ab85-0e8eb61422e2" containerID="4a541af3fa0c88410e322db126e7852014165b17d79a5029bf58c38739650f88" exitCode=2 Dec 05 08:39:48 crc kubenswrapper[4876]: I1205 08:39:48.698795 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b29eddc0-b7dd-4390-ab85-0e8eb61422e2","Type":"ContainerDied","Data":"4a541af3fa0c88410e322db126e7852014165b17d79a5029bf58c38739650f88"} Dec 05 08:39:48 crc kubenswrapper[4876]: I1205 08:39:48.830233 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 08:39:48 crc kubenswrapper[4876]: I1205 08:39:48.909238 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klg97\" (UniqueName: \"kubernetes.io/projected/b29eddc0-b7dd-4390-ab85-0e8eb61422e2-kube-api-access-klg97\") pod \"b29eddc0-b7dd-4390-ab85-0e8eb61422e2\" (UID: \"b29eddc0-b7dd-4390-ab85-0e8eb61422e2\") " Dec 05 08:39:48 crc kubenswrapper[4876]: I1205 08:39:48.915956 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b29eddc0-b7dd-4390-ab85-0e8eb61422e2-kube-api-access-klg97" (OuterVolumeSpecName: "kube-api-access-klg97") pod "b29eddc0-b7dd-4390-ab85-0e8eb61422e2" (UID: "b29eddc0-b7dd-4390-ab85-0e8eb61422e2"). InnerVolumeSpecName "kube-api-access-klg97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.012157 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klg97\" (UniqueName: \"kubernetes.io/projected/b29eddc0-b7dd-4390-ab85-0e8eb61422e2-kube-api-access-klg97\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.711449 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b29eddc0-b7dd-4390-ab85-0e8eb61422e2","Type":"ContainerDied","Data":"75e0ed124f7c0f5d1179aa1d7f39781a3df329d3eb9d1b21201010653d86d0eb"} Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.711501 4876 scope.go:117] "RemoveContainer" containerID="4a541af3fa0c88410e322db126e7852014165b17d79a5029bf58c38739650f88" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.711502 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.742268 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.753483 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.766708 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:39:49 crc kubenswrapper[4876]: E1205 08:39:49.767210 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bc9681-6da3-4e8b-8a88-4940f7b359de" containerName="dnsmasq-dns" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.767230 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bc9681-6da3-4e8b-8a88-4940f7b359de" containerName="dnsmasq-dns" Dec 05 08:39:49 crc kubenswrapper[4876]: E1205 08:39:49.767260 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bc9681-6da3-4e8b-8a88-4940f7b359de" containerName="init" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.767266 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bc9681-6da3-4e8b-8a88-4940f7b359de" containerName="init" Dec 05 08:39:49 crc kubenswrapper[4876]: E1205 08:39:49.767278 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b29eddc0-b7dd-4390-ab85-0e8eb61422e2" containerName="kube-state-metrics" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.767284 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b29eddc0-b7dd-4390-ab85-0e8eb61422e2" containerName="kube-state-metrics" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.767467 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="14bc9681-6da3-4e8b-8a88-4940f7b359de" containerName="dnsmasq-dns" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.767478 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b29eddc0-b7dd-4390-ab85-0e8eb61422e2" containerName="kube-state-metrics" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.768271 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.770354 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.770594 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.776660 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.825828 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7ldc\" (UniqueName: \"kubernetes.io/projected/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-api-access-d7ldc\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.825877 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.826001 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.826138 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.836259 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b29eddc0-b7dd-4390-ab85-0e8eb61422e2" path="/var/lib/kubelet/pods/b29eddc0-b7dd-4390-ab85-0e8eb61422e2/volumes" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.927513 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7ldc\" (UniqueName: \"kubernetes.io/projected/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-api-access-d7ldc\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.927768 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.927949 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.928136 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.932478 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.933971 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.935446 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:49 crc kubenswrapper[4876]: I1205 08:39:49.943646 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7ldc\" (UniqueName: \"kubernetes.io/projected/1c06bc09-5dd5-4d2c-b572-078ea2a4d037-kube-api-access-d7ldc\") pod \"kube-state-metrics-0\" (UID: \"1c06bc09-5dd5-4d2c-b572-078ea2a4d037\") " pod="openstack/kube-state-metrics-0" Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.054492 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.054756 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="ceilometer-central-agent" containerID="cri-o://11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd" gracePeriod=30 Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.054863 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="proxy-httpd" containerID="cri-o://a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286" gracePeriod=30 Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.054926 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="sg-core" containerID="cri-o://230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5" gracePeriod=30 Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.054969 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="ceilometer-notification-agent" containerID="cri-o://fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056" gracePeriod=30 Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.126510 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.578431 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.725867 4876 generic.go:334] "Generic (PLEG): container finished" podID="239ff295-f4fb-4d94-b893-76c436b41015" containerID="a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286" exitCode=0 Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.725917 4876 generic.go:334] "Generic (PLEG): container finished" podID="239ff295-f4fb-4d94-b893-76c436b41015" containerID="230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5" exitCode=2 Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.725931 4876 generic.go:334] "Generic (PLEG): container finished" podID="239ff295-f4fb-4d94-b893-76c436b41015" containerID="11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd" exitCode=0 Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.725978 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerDied","Data":"a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286"} Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.726008 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerDied","Data":"230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5"} Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.726021 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerDied","Data":"11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd"} Dec 05 08:39:50 crc kubenswrapper[4876]: I1205 08:39:50.730314 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1c06bc09-5dd5-4d2c-b572-078ea2a4d037","Type":"ContainerStarted","Data":"c108427f7f7eda063af31e824017c5e0301b52af1e28890d358d356cab4fc40e"} Dec 05 08:39:51 crc kubenswrapper[4876]: I1205 08:39:51.741953 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1c06bc09-5dd5-4d2c-b572-078ea2a4d037","Type":"ContainerStarted","Data":"601d8712566de926a3bcc84d27b21d2cf26d42078eb22c17f192b5095aebd4d4"} Dec 05 08:39:51 crc kubenswrapper[4876]: I1205 08:39:51.743316 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 08:39:51 crc kubenswrapper[4876]: I1205 08:39:51.760616 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.324752294 podStartE2EDuration="2.760600548s" podCreationTimestamp="2025-12-05 08:39:49 +0000 UTC" firstStartedPulling="2025-12-05 08:39:50.581126332 +0000 UTC m=+1455.069790954" lastFinishedPulling="2025-12-05 08:39:51.016974586 +0000 UTC m=+1455.505639208" observedRunningTime="2025-12-05 08:39:51.759743365 +0000 UTC m=+1456.248407997" watchObservedRunningTime="2025-12-05 08:39:51.760600548 +0000 UTC m=+1456.249265170" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.445233 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.474068 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-log-httpd\") pod \"239ff295-f4fb-4d94-b893-76c436b41015\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.474120 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-config-data\") pod \"239ff295-f4fb-4d94-b893-76c436b41015\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.474173 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-combined-ca-bundle\") pod \"239ff295-f4fb-4d94-b893-76c436b41015\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.474255 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-scripts\") pod \"239ff295-f4fb-4d94-b893-76c436b41015\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.474336 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-run-httpd\") pod \"239ff295-f4fb-4d94-b893-76c436b41015\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.474375 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25dbf\" (UniqueName: \"kubernetes.io/projected/239ff295-f4fb-4d94-b893-76c436b41015-kube-api-access-25dbf\") pod \"239ff295-f4fb-4d94-b893-76c436b41015\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.474421 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-sg-core-conf-yaml\") pod \"239ff295-f4fb-4d94-b893-76c436b41015\" (UID: \"239ff295-f4fb-4d94-b893-76c436b41015\") " Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.474923 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "239ff295-f4fb-4d94-b893-76c436b41015" (UID: "239ff295-f4fb-4d94-b893-76c436b41015"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.475292 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "239ff295-f4fb-4d94-b893-76c436b41015" (UID: "239ff295-f4fb-4d94-b893-76c436b41015"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.486966 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-scripts" (OuterVolumeSpecName: "scripts") pod "239ff295-f4fb-4d94-b893-76c436b41015" (UID: "239ff295-f4fb-4d94-b893-76c436b41015"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.487319 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239ff295-f4fb-4d94-b893-76c436b41015-kube-api-access-25dbf" (OuterVolumeSpecName: "kube-api-access-25dbf") pod "239ff295-f4fb-4d94-b893-76c436b41015" (UID: "239ff295-f4fb-4d94-b893-76c436b41015"). InnerVolumeSpecName "kube-api-access-25dbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.515152 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "239ff295-f4fb-4d94-b893-76c436b41015" (UID: "239ff295-f4fb-4d94-b893-76c436b41015"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.566640 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "239ff295-f4fb-4d94-b893-76c436b41015" (UID: "239ff295-f4fb-4d94-b893-76c436b41015"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.575553 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.575585 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25dbf\" (UniqueName: \"kubernetes.io/projected/239ff295-f4fb-4d94-b893-76c436b41015-kube-api-access-25dbf\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.575599 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.575609 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/239ff295-f4fb-4d94-b893-76c436b41015-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.575620 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.575632 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.610497 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-config-data" (OuterVolumeSpecName: "config-data") pod "239ff295-f4fb-4d94-b893-76c436b41015" (UID: "239ff295-f4fb-4d94-b893-76c436b41015"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.677315 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/239ff295-f4fb-4d94-b893-76c436b41015-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.778099 4876 generic.go:334] "Generic (PLEG): container finished" podID="239ff295-f4fb-4d94-b893-76c436b41015" containerID="fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056" exitCode=0 Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.778150 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerDied","Data":"fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056"} Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.778169 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.778188 4876 scope.go:117] "RemoveContainer" containerID="a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.778176 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"239ff295-f4fb-4d94-b893-76c436b41015","Type":"ContainerDied","Data":"475ad07415119597e38d9c988d756565b5cb1792beb2885462b0951907bf4751"} Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.813693 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.815075 4876 scope.go:117] "RemoveContainer" containerID="230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.825781 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841063 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:54 crc kubenswrapper[4876]: E1205 08:39:54.841450 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="ceilometer-central-agent" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841468 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="ceilometer-central-agent" Dec 05 08:39:54 crc kubenswrapper[4876]: E1205 08:39:54.841487 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="proxy-httpd" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841493 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="proxy-httpd" Dec 05 08:39:54 crc kubenswrapper[4876]: E1205 08:39:54.841508 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="sg-core" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841514 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="sg-core" Dec 05 08:39:54 crc kubenswrapper[4876]: E1205 08:39:54.841529 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="ceilometer-notification-agent" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841535 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="ceilometer-notification-agent" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841715 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="ceilometer-central-agent" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841741 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="proxy-httpd" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841749 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="sg-core" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.841760 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="239ff295-f4fb-4d94-b893-76c436b41015" containerName="ceilometer-notification-agent" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.843418 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.846078 4876 scope.go:117] "RemoveContainer" containerID="fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.849463 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.849727 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.849952 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.865035 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.881256 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.881368 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-scripts\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.881479 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-log-httpd\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.881516 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z66cf\" (UniqueName: \"kubernetes.io/projected/d065d71e-e44d-426d-87e9-5924a1ed5784-kube-api-access-z66cf\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.881574 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-run-httpd\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.881597 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-config-data\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.881645 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.881709 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.888222 4876 scope.go:117] "RemoveContainer" containerID="11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.911981 4876 scope.go:117] "RemoveContainer" containerID="a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286" Dec 05 08:39:54 crc kubenswrapper[4876]: E1205 08:39:54.912447 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286\": container with ID starting with a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286 not found: ID does not exist" containerID="a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.912489 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286"} err="failed to get container status \"a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286\": rpc error: code = NotFound desc = could not find container \"a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286\": container with ID starting with a84c536a10ee24fde1671cd722dcc81ecd54259f299314530d53f299c2fe5286 not found: ID does not exist" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.912515 4876 scope.go:117] "RemoveContainer" containerID="230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5" Dec 05 08:39:54 crc kubenswrapper[4876]: E1205 08:39:54.912853 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5\": container with ID starting with 230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5 not found: ID does not exist" containerID="230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.912964 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5"} err="failed to get container status \"230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5\": rpc error: code = NotFound desc = could not find container \"230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5\": container with ID starting with 230f847c9859f804c0f489f5ecc2d306595ca0015b010dbb5eaff9f08462edd5 not found: ID does not exist" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.912998 4876 scope.go:117] "RemoveContainer" containerID="fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056" Dec 05 08:39:54 crc kubenswrapper[4876]: E1205 08:39:54.913302 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056\": container with ID starting with fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056 not found: ID does not exist" containerID="fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.913341 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056"} err="failed to get container status \"fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056\": rpc error: code = NotFound desc = could not find container \"fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056\": container with ID starting with fa238714e4c8e1ef67a5dab3ef46fbf34dc9db265d27ab39daecd5e565508056 not found: ID does not exist" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.913366 4876 scope.go:117] "RemoveContainer" containerID="11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd" Dec 05 08:39:54 crc kubenswrapper[4876]: E1205 08:39:54.913636 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd\": container with ID starting with 11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd not found: ID does not exist" containerID="11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.913663 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd"} err="failed to get container status \"11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd\": rpc error: code = NotFound desc = could not find container \"11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd\": container with ID starting with 11a4ae9168da6c10eab4a0c59b6eaad2f38e376de8dcccf12e1e6b49e8a38abd not found: ID does not exist" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.988774 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-scripts\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.988994 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-log-httpd\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.989035 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z66cf\" (UniqueName: \"kubernetes.io/projected/d065d71e-e44d-426d-87e9-5924a1ed5784-kube-api-access-z66cf\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.989126 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-run-httpd\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.989163 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-config-data\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.989236 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.989355 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.989440 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.990089 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-run-httpd\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.990236 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-log-httpd\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.993757 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:54 crc kubenswrapper[4876]: I1205 08:39:54.994856 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-scripts\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:55 crc kubenswrapper[4876]: I1205 08:39:55.001314 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-config-data\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:55 crc kubenswrapper[4876]: I1205 08:39:55.004547 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:55 crc kubenswrapper[4876]: I1205 08:39:55.011757 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z66cf\" (UniqueName: \"kubernetes.io/projected/d065d71e-e44d-426d-87e9-5924a1ed5784-kube-api-access-z66cf\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:55 crc kubenswrapper[4876]: I1205 08:39:55.016910 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " pod="openstack/ceilometer-0" Dec 05 08:39:55 crc kubenswrapper[4876]: I1205 08:39:55.169710 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 08:39:55 crc kubenswrapper[4876]: I1205 08:39:55.639596 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 08:39:55 crc kubenswrapper[4876]: I1205 08:39:55.789794 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerStarted","Data":"39c72c126e9850f6c76e9e6df2f39cd7063bd3f7a876509778d56064d237cb9f"} Dec 05 08:39:55 crc kubenswrapper[4876]: I1205 08:39:55.841394 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239ff295-f4fb-4d94-b893-76c436b41015" path="/var/lib/kubelet/pods/239ff295-f4fb-4d94-b893-76c436b41015/volumes" Dec 05 08:39:56 crc kubenswrapper[4876]: I1205 08:39:56.804109 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerStarted","Data":"617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8"} Dec 05 08:39:57 crc kubenswrapper[4876]: I1205 08:39:57.796576 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:39:57 crc kubenswrapper[4876]: I1205 08:39:57.819444 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerStarted","Data":"68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765"} Dec 05 08:39:57 crc kubenswrapper[4876]: I1205 08:39:57.819491 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerStarted","Data":"63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326"} Dec 05 08:39:58 crc kubenswrapper[4876]: I1205 08:39:58.866120 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:39:59 crc kubenswrapper[4876]: I1205 08:39:59.865105 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerStarted","Data":"e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3"} Dec 05 08:39:59 crc kubenswrapper[4876]: I1205 08:39:59.865671 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 08:39:59 crc kubenswrapper[4876]: I1205 08:39:59.891748 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.838621173 podStartE2EDuration="5.891730021s" podCreationTimestamp="2025-12-05 08:39:54 +0000 UTC" firstStartedPulling="2025-12-05 08:39:55.64267593 +0000 UTC m=+1460.131340552" lastFinishedPulling="2025-12-05 08:39:58.695784778 +0000 UTC m=+1463.184449400" observedRunningTime="2025-12-05 08:39:59.887323411 +0000 UTC m=+1464.375988033" watchObservedRunningTime="2025-12-05 08:39:59.891730021 +0000 UTC m=+1464.380394653" Dec 05 08:40:00 crc kubenswrapper[4876]: I1205 08:40:00.145627 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 08:40:02 crc kubenswrapper[4876]: I1205 08:40:02.497687 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerName="rabbitmq" containerID="cri-o://6cc496e7c01898528567beed9d06b8072750b24382edf8aea2e624e6163bbd2b" gracePeriod=604796 Dec 05 08:40:03 crc kubenswrapper[4876]: I1205 08:40:03.204103 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerName="rabbitmq" containerID="cri-o://a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6" gracePeriod=604796 Dec 05 08:40:07 crc kubenswrapper[4876]: I1205 08:40:07.438188 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 05 08:40:08 crc kubenswrapper[4876]: I1205 08:40:08.089014 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 05 08:40:08 crc kubenswrapper[4876]: I1205 08:40:08.213721 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:40:08 crc kubenswrapper[4876]: I1205 08:40:08.213806 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:40:08 crc kubenswrapper[4876]: I1205 08:40:08.971215 4876 generic.go:334] "Generic (PLEG): container finished" podID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerID="6cc496e7c01898528567beed9d06b8072750b24382edf8aea2e624e6163bbd2b" exitCode=0 Dec 05 08:40:08 crc kubenswrapper[4876]: I1205 08:40:08.971349 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f","Type":"ContainerDied","Data":"6cc496e7c01898528567beed9d06b8072750b24382edf8aea2e624e6163bbd2b"} Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.208726 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.372855 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkq9s\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-kube-api-access-qkq9s\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373058 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-config-data\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373137 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-pod-info\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373179 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-erlang-cookie-secret\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373214 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-plugins-conf\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373277 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-plugins\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373307 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373345 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-erlang-cookie\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373370 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-server-conf\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373412 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-confd\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.373434 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-tls\") pod \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\" (UID: \"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.374263 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.374385 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.374630 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.380403 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-pod-info" (OuterVolumeSpecName: "pod-info") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.382756 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.382814 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-kube-api-access-qkq9s" (OuterVolumeSpecName: "kube-api-access-qkq9s") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "kube-api-access-qkq9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.383704 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.384095 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.408870 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-config-data" (OuterVolumeSpecName: "config-data") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.473569 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-server-conf" (OuterVolumeSpecName: "server-conf") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.475852 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.475921 4876 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.475934 4876 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.475945 4876 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.475956 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.476023 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.476083 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.476096 4876 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.476107 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.476117 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkq9s\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-kube-api-access-qkq9s\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.499959 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.545470 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" (UID: "7edf4ad5-0a62-4a2d-af4a-aecf59072c8f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.578094 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.578134 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.834772 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.889721 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-server-conf\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.889808 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rljh\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-kube-api-access-4rljh\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.889849 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-tls\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.889914 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-confd\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.889967 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-erlang-cookie\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.890011 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/45b18655-19ec-43f4-a779-b8f72bc77aa2-erlang-cookie-secret\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.890038 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/45b18655-19ec-43f4-a779-b8f72bc77aa2-pod-info\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.890068 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-plugins\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.890104 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-config-data\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.890144 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.890167 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-plugins-conf\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.890494 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.891039 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.894937 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.895759 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.909943 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-kube-api-access-4rljh" (OuterVolumeSpecName: "kube-api-access-4rljh") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "kube-api-access-4rljh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.910301 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.911298 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.921485 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45b18655-19ec-43f4-a779-b8f72bc77aa2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.929243 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/45b18655-19ec-43f4-a779-b8f72bc77aa2-pod-info" (OuterVolumeSpecName: "pod-info") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.938917 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-config-data" (OuterVolumeSpecName: "config-data") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.992310 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-server-conf" (OuterVolumeSpecName: "server-conf") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.992479 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-server-conf\") pod \"45b18655-19ec-43f4-a779-b8f72bc77aa2\" (UID: \"45b18655-19ec-43f4-a779-b8f72bc77aa2\") " Dec 05 08:40:09 crc kubenswrapper[4876]: W1205 08:40:09.993277 4876 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/45b18655-19ec-43f4-a779-b8f72bc77aa2/volumes/kubernetes.io~configmap/server-conf Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.993311 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-server-conf" (OuterVolumeSpecName: "server-conf") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995033 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995057 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995078 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995088 4876 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995097 4876 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/45b18655-19ec-43f4-a779-b8f72bc77aa2-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995106 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rljh\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-kube-api-access-4rljh\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995116 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995124 4876 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/45b18655-19ec-43f4-a779-b8f72bc77aa2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:09 crc kubenswrapper[4876]: I1205 08:40:09.995131 4876 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/45b18655-19ec-43f4-a779-b8f72bc77aa2-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.002650 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7edf4ad5-0a62-4a2d-af4a-aecf59072c8f","Type":"ContainerDied","Data":"ac3247a496e8447132b39c1b5aa4f3e18d81a65ca345b38845a7e18d7235214d"} Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.002707 4876 scope.go:117] "RemoveContainer" containerID="6cc496e7c01898528567beed9d06b8072750b24382edf8aea2e624e6163bbd2b" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.002719 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.015674 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.016716 4876 generic.go:334] "Generic (PLEG): container finished" podID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerID="a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6" exitCode=0 Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.016759 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45b18655-19ec-43f4-a779-b8f72bc77aa2","Type":"ContainerDied","Data":"a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6"} Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.016773 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.016789 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"45b18655-19ec-43f4-a779-b8f72bc77aa2","Type":"ContainerDied","Data":"56c55451e074fb1f9579db7475e31ac8e5c00ad401cb11503154d025e630feaf"} Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.025923 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "45b18655-19ec-43f4-a779-b8f72bc77aa2" (UID: "45b18655-19ec-43f4-a779-b8f72bc77aa2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.045447 4876 scope.go:117] "RemoveContainer" containerID="59bcde78d239d2ec5b2c1b6c42bfe979ce595826cff0ba2f07bdfb82de0dac70" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.047336 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.058335 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.065749 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:40:10 crc kubenswrapper[4876]: E1205 08:40:10.066176 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerName="setup-container" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.066198 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerName="setup-container" Dec 05 08:40:10 crc kubenswrapper[4876]: E1205 08:40:10.066215 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerName="rabbitmq" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.066224 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerName="rabbitmq" Dec 05 08:40:10 crc kubenswrapper[4876]: E1205 08:40:10.066257 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerName="rabbitmq" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.066265 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerName="rabbitmq" Dec 05 08:40:10 crc kubenswrapper[4876]: E1205 08:40:10.066290 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerName="setup-container" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.066298 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerName="setup-container" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.066507 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" containerName="rabbitmq" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.066535 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" containerName="rabbitmq" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.067872 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.073702 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h2wdn" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.073895 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.074015 4876 scope.go:117] "RemoveContainer" containerID="a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.074044 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.074181 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.074204 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.074489 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.074491 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.087620 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.096618 4876 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/45b18655-19ec-43f4-a779-b8f72bc77aa2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.096643 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.105393 4876 scope.go:117] "RemoveContainer" containerID="f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.142109 4876 scope.go:117] "RemoveContainer" containerID="a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6" Dec 05 08:40:10 crc kubenswrapper[4876]: E1205 08:40:10.142677 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6\": container with ID starting with a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6 not found: ID does not exist" containerID="a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.142748 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6"} err="failed to get container status \"a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6\": rpc error: code = NotFound desc = could not find container \"a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6\": container with ID starting with a2771818101c0f0f2b8a5a0da24da98e7388b2873764782852b8e373e8287bf6 not found: ID does not exist" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.142779 4876 scope.go:117] "RemoveContainer" containerID="f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369" Dec 05 08:40:10 crc kubenswrapper[4876]: E1205 08:40:10.143282 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369\": container with ID starting with f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369 not found: ID does not exist" containerID="f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.143338 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369"} err="failed to get container status \"f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369\": rpc error: code = NotFound desc = could not find container \"f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369\": container with ID starting with f5dc1cdf01f2cf12ee92685b981571b41fda10e76b94e863036c206f93e93369 not found: ID does not exist" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.198490 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.198544 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/692469d9-b150-4d67-b412-ba8760598f8b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.198579 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-config-data\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.198609 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckjc6\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-kube-api-access-ckjc6\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.198651 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.198732 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.198836 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.198980 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.199014 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.199149 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/692469d9-b150-4d67-b412-ba8760598f8b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.199235 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.300857 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.300927 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.300983 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/692469d9-b150-4d67-b412-ba8760598f8b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.301017 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.301053 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.301077 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/692469d9-b150-4d67-b412-ba8760598f8b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.301099 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-config-data\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.301121 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckjc6\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-kube-api-access-ckjc6\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.301140 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.301160 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.301188 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.302033 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.302133 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.302169 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.302234 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.302324 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.302526 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/692469d9-b150-4d67-b412-ba8760598f8b-config-data\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.312556 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/692469d9-b150-4d67-b412-ba8760598f8b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.313704 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/692469d9-b150-4d67-b412-ba8760598f8b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.316830 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.319355 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.335774 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckjc6\" (UniqueName: \"kubernetes.io/projected/692469d9-b150-4d67-b412-ba8760598f8b-kube-api-access-ckjc6\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.377313 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"692469d9-b150-4d67-b412-ba8760598f8b\") " pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.388359 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.403972 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.425787 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.444001 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.445615 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.448232 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mmjr5" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.449082 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.449244 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.449515 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.450070 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.450142 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.454640 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.493977 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508048 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508098 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508159 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508192 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508262 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508324 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6ef895f6-3d13-4980-98ef-9111438fe1ec-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508384 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508412 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508460 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508477 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g49h6\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-kube-api-access-g49h6\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.508621 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6ef895f6-3d13-4980-98ef-9111438fe1ec-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.611457 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.611531 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.611597 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6ef895f6-3d13-4980-98ef-9111438fe1ec-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.611621 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.611642 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.611669 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.611690 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g49h6\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-kube-api-access-g49h6\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.612404 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6ef895f6-3d13-4980-98ef-9111438fe1ec-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.612506 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.612544 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.612634 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.613172 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.614149 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.614332 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.614647 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.614759 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.615463 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6ef895f6-3d13-4980-98ef-9111438fe1ec-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.621498 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6ef895f6-3d13-4980-98ef-9111438fe1ec-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.621503 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.632527 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6ef895f6-3d13-4980-98ef-9111438fe1ec-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.635577 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.636483 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g49h6\" (UniqueName: \"kubernetes.io/projected/6ef895f6-3d13-4980-98ef-9111438fe1ec-kube-api-access-g49h6\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.645858 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6ef895f6-3d13-4980-98ef-9111438fe1ec\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.772260 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:10 crc kubenswrapper[4876]: I1205 08:40:10.938302 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 08:40:11 crc kubenswrapper[4876]: I1205 08:40:11.033351 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"692469d9-b150-4d67-b412-ba8760598f8b","Type":"ContainerStarted","Data":"b3a068462257e65849b84e9739f9450abb247984c6951fea9977315b78e579e5"} Dec 05 08:40:11 crc kubenswrapper[4876]: I1205 08:40:11.273174 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 08:40:11 crc kubenswrapper[4876]: W1205 08:40:11.287363 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ef895f6_3d13_4980_98ef_9111438fe1ec.slice/crio-5571fd486543251f30acfcd5e7fbfc9b391724dce426e0e838b6adb362e9d437 WatchSource:0}: Error finding container 5571fd486543251f30acfcd5e7fbfc9b391724dce426e0e838b6adb362e9d437: Status 404 returned error can't find the container with id 5571fd486543251f30acfcd5e7fbfc9b391724dce426e0e838b6adb362e9d437 Dec 05 08:40:11 crc kubenswrapper[4876]: I1205 08:40:11.838106 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45b18655-19ec-43f4-a779-b8f72bc77aa2" path="/var/lib/kubelet/pods/45b18655-19ec-43f4-a779-b8f72bc77aa2/volumes" Dec 05 08:40:11 crc kubenswrapper[4876]: I1205 08:40:11.839992 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7edf4ad5-0a62-4a2d-af4a-aecf59072c8f" path="/var/lib/kubelet/pods/7edf4ad5-0a62-4a2d-af4a-aecf59072c8f/volumes" Dec 05 08:40:12 crc kubenswrapper[4876]: I1205 08:40:12.047814 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6ef895f6-3d13-4980-98ef-9111438fe1ec","Type":"ContainerStarted","Data":"5571fd486543251f30acfcd5e7fbfc9b391724dce426e0e838b6adb362e9d437"} Dec 05 08:40:13 crc kubenswrapper[4876]: I1205 08:40:13.064027 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"692469d9-b150-4d67-b412-ba8760598f8b","Type":"ContainerStarted","Data":"a6ff9c681250d1426f57c1ba83382bd24bc687513b76ea7c2e3db6bc5c50f907"} Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.080034 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6ef895f6-3d13-4980-98ef-9111438fe1ec","Type":"ContainerStarted","Data":"2a3abf57515fd9df86edcaf44fe6c62332e0b96fc4c3e93ec0c054e366e7b26f"} Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.206829 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-g797c"] Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.208575 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.219961 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.229007 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-g797c"] Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.301980 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.302035 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-config\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.302061 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.302375 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7vgm\" (UniqueName: \"kubernetes.io/projected/cc0b0027-b82a-47a8-900c-6de076094417-kube-api-access-b7vgm\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.302426 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.302593 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-svc\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.302628 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.404131 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7vgm\" (UniqueName: \"kubernetes.io/projected/cc0b0027-b82a-47a8-900c-6de076094417-kube-api-access-b7vgm\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.404202 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.404276 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-svc\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.404305 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.404340 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.404361 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-config\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.404381 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.405271 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.405322 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.406061 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.406073 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.406658 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-svc\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.406682 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-config\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.427505 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7vgm\" (UniqueName: \"kubernetes.io/projected/cc0b0027-b82a-47a8-900c-6de076094417-kube-api-access-b7vgm\") pod \"dnsmasq-dns-594cb89c79-g797c\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.537400 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:14 crc kubenswrapper[4876]: I1205 08:40:14.988716 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-g797c"] Dec 05 08:40:14 crc kubenswrapper[4876]: W1205 08:40:14.991325 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc0b0027_b82a_47a8_900c_6de076094417.slice/crio-05d39026a0ecc9a3f90678a38e9e846d262ae70fb009149b1c6d2e2fab0710b6 WatchSource:0}: Error finding container 05d39026a0ecc9a3f90678a38e9e846d262ae70fb009149b1c6d2e2fab0710b6: Status 404 returned error can't find the container with id 05d39026a0ecc9a3f90678a38e9e846d262ae70fb009149b1c6d2e2fab0710b6 Dec 05 08:40:15 crc kubenswrapper[4876]: I1205 08:40:15.094361 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-g797c" event={"ID":"cc0b0027-b82a-47a8-900c-6de076094417","Type":"ContainerStarted","Data":"05d39026a0ecc9a3f90678a38e9e846d262ae70fb009149b1c6d2e2fab0710b6"} Dec 05 08:40:16 crc kubenswrapper[4876]: I1205 08:40:16.105398 4876 generic.go:334] "Generic (PLEG): container finished" podID="cc0b0027-b82a-47a8-900c-6de076094417" containerID="e9a6c2f76bea61e39105b3b1bafb425bb73d212aaf3a7ac0b6a0728afe29eb71" exitCode=0 Dec 05 08:40:16 crc kubenswrapper[4876]: I1205 08:40:16.105483 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-g797c" event={"ID":"cc0b0027-b82a-47a8-900c-6de076094417","Type":"ContainerDied","Data":"e9a6c2f76bea61e39105b3b1bafb425bb73d212aaf3a7ac0b6a0728afe29eb71"} Dec 05 08:40:17 crc kubenswrapper[4876]: I1205 08:40:17.118964 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-g797c" event={"ID":"cc0b0027-b82a-47a8-900c-6de076094417","Type":"ContainerStarted","Data":"58b64f41cb01a3ac1c41bcfdc26018a761cfa3739bbb5910499c05cbc8a51689"} Dec 05 08:40:17 crc kubenswrapper[4876]: I1205 08:40:17.119155 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:17 crc kubenswrapper[4876]: I1205 08:40:17.147228 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-594cb89c79-g797c" podStartSLOduration=3.147206206 podStartE2EDuration="3.147206206s" podCreationTimestamp="2025-12-05 08:40:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:40:17.141939944 +0000 UTC m=+1481.630604596" watchObservedRunningTime="2025-12-05 08:40:17.147206206 +0000 UTC m=+1481.635870838" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.540270 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.609836 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-tmh27"] Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.610177 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" podUID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" containerName="dnsmasq-dns" containerID="cri-o://6c3ff6b383912f24151094bdbc91f3b7716fc94c398919c79a9b60cf6e5a50ea" gracePeriod=10 Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.789292 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-dcb46"] Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.797228 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.807137 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-dcb46"] Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.941523 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-config\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.941623 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.941748 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.941844 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsf99\" (UniqueName: \"kubernetes.io/projected/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-kube-api-access-bsf99\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.941925 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.942006 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:24 crc kubenswrapper[4876]: I1205 08:40:24.942110 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.044392 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-config\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.044488 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.044515 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.044541 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsf99\" (UniqueName: \"kubernetes.io/projected/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-kube-api-access-bsf99\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.044559 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.044582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.044620 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.045614 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.045968 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.046226 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.046688 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-config\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.046687 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.047065 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.063996 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsf99\" (UniqueName: \"kubernetes.io/projected/ece6a347-25cd-41e9-89e0-f20c1fdb9eba-kube-api-access-bsf99\") pod \"dnsmasq-dns-5596c69fcc-dcb46\" (UID: \"ece6a347-25cd-41e9-89e0-f20c1fdb9eba\") " pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.130954 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.180013 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.239157 4876 generic.go:334] "Generic (PLEG): container finished" podID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" containerID="6c3ff6b383912f24151094bdbc91f3b7716fc94c398919c79a9b60cf6e5a50ea" exitCode=0 Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.239252 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" event={"ID":"2db748c5-cab9-4494-8ba0-ff71f7b8b252","Type":"ContainerDied","Data":"6c3ff6b383912f24151094bdbc91f3b7716fc94c398919c79a9b60cf6e5a50ea"} Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.239623 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" event={"ID":"2db748c5-cab9-4494-8ba0-ff71f7b8b252","Type":"ContainerDied","Data":"a282cc6589c82d98e6a415be0894ff9b3401e543bbd8f68c3dbc7e908bdb55bc"} Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.239850 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a282cc6589c82d98e6a415be0894ff9b3401e543bbd8f68c3dbc7e908bdb55bc" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.260951 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.452604 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-sb\") pod \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.452773 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-svc\") pod \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.452847 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24lt7\" (UniqueName: \"kubernetes.io/projected/2db748c5-cab9-4494-8ba0-ff71f7b8b252-kube-api-access-24lt7\") pod \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.452995 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-nb\") pod \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.453046 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-swift-storage-0\") pod \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.453101 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-config\") pod \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\" (UID: \"2db748c5-cab9-4494-8ba0-ff71f7b8b252\") " Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.458964 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2db748c5-cab9-4494-8ba0-ff71f7b8b252-kube-api-access-24lt7" (OuterVolumeSpecName: "kube-api-access-24lt7") pod "2db748c5-cab9-4494-8ba0-ff71f7b8b252" (UID: "2db748c5-cab9-4494-8ba0-ff71f7b8b252"). InnerVolumeSpecName "kube-api-access-24lt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.517400 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2db748c5-cab9-4494-8ba0-ff71f7b8b252" (UID: "2db748c5-cab9-4494-8ba0-ff71f7b8b252"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.523643 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2db748c5-cab9-4494-8ba0-ff71f7b8b252" (UID: "2db748c5-cab9-4494-8ba0-ff71f7b8b252"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.529037 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2db748c5-cab9-4494-8ba0-ff71f7b8b252" (UID: "2db748c5-cab9-4494-8ba0-ff71f7b8b252"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.530391 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-config" (OuterVolumeSpecName: "config") pod "2db748c5-cab9-4494-8ba0-ff71f7b8b252" (UID: "2db748c5-cab9-4494-8ba0-ff71f7b8b252"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.535272 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2db748c5-cab9-4494-8ba0-ff71f7b8b252" (UID: "2db748c5-cab9-4494-8ba0-ff71f7b8b252"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.556300 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.556366 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24lt7\" (UniqueName: \"kubernetes.io/projected/2db748c5-cab9-4494-8ba0-ff71f7b8b252-kube-api-access-24lt7\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.556389 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.556410 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.556426 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:25 crc kubenswrapper[4876]: I1205 08:40:25.556440 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2db748c5-cab9-4494-8ba0-ff71f7b8b252-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:26 crc kubenswrapper[4876]: I1205 08:40:26.210374 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-dcb46"] Dec 05 08:40:26 crc kubenswrapper[4876]: I1205 08:40:26.259795 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" event={"ID":"ece6a347-25cd-41e9-89e0-f20c1fdb9eba","Type":"ContainerStarted","Data":"16b4cea5845618b3949edd8f78b2fada70f0f76c239cae614d153cd93dadcb73"} Dec 05 08:40:26 crc kubenswrapper[4876]: I1205 08:40:26.259820 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-tmh27" Dec 05 08:40:26 crc kubenswrapper[4876]: I1205 08:40:26.322386 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-tmh27"] Dec 05 08:40:26 crc kubenswrapper[4876]: I1205 08:40:26.333825 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-tmh27"] Dec 05 08:40:27 crc kubenswrapper[4876]: I1205 08:40:27.268841 4876 generic.go:334] "Generic (PLEG): container finished" podID="ece6a347-25cd-41e9-89e0-f20c1fdb9eba" containerID="bdb936cf15a02d70b9530944bb0079014eeb7a93c621ec4de64213b8252c1e79" exitCode=0 Dec 05 08:40:27 crc kubenswrapper[4876]: I1205 08:40:27.268916 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" event={"ID":"ece6a347-25cd-41e9-89e0-f20c1fdb9eba","Type":"ContainerDied","Data":"bdb936cf15a02d70b9530944bb0079014eeb7a93c621ec4de64213b8252c1e79"} Dec 05 08:40:27 crc kubenswrapper[4876]: I1205 08:40:27.833672 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" path="/var/lib/kubelet/pods/2db748c5-cab9-4494-8ba0-ff71f7b8b252/volumes" Dec 05 08:40:28 crc kubenswrapper[4876]: I1205 08:40:28.279961 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" event={"ID":"ece6a347-25cd-41e9-89e0-f20c1fdb9eba","Type":"ContainerStarted","Data":"9564f4aec3d419d836f767aa4f6018a789831f0dd5eeb88209b0368058cd9e87"} Dec 05 08:40:28 crc kubenswrapper[4876]: I1205 08:40:28.281074 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:28 crc kubenswrapper[4876]: I1205 08:40:28.306820 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" podStartSLOduration=4.30680033 podStartE2EDuration="4.30680033s" podCreationTimestamp="2025-12-05 08:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:40:28.300817263 +0000 UTC m=+1492.789481895" watchObservedRunningTime="2025-12-05 08:40:28.30680033 +0000 UTC m=+1492.795464962" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.132892 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5596c69fcc-dcb46" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.235051 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-g797c"] Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.235503 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-594cb89c79-g797c" podUID="cc0b0027-b82a-47a8-900c-6de076094417" containerName="dnsmasq-dns" containerID="cri-o://58b64f41cb01a3ac1c41bcfdc26018a761cfa3739bbb5910499c05cbc8a51689" gracePeriod=10 Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.367147 4876 generic.go:334] "Generic (PLEG): container finished" podID="cc0b0027-b82a-47a8-900c-6de076094417" containerID="58b64f41cb01a3ac1c41bcfdc26018a761cfa3739bbb5910499c05cbc8a51689" exitCode=0 Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.367221 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-g797c" event={"ID":"cc0b0027-b82a-47a8-900c-6de076094417","Type":"ContainerDied","Data":"58b64f41cb01a3ac1c41bcfdc26018a761cfa3739bbb5910499c05cbc8a51689"} Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.770297 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.921552 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-swift-storage-0\") pod \"cc0b0027-b82a-47a8-900c-6de076094417\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.921635 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7vgm\" (UniqueName: \"kubernetes.io/projected/cc0b0027-b82a-47a8-900c-6de076094417-kube-api-access-b7vgm\") pod \"cc0b0027-b82a-47a8-900c-6de076094417\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.921662 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-nb\") pod \"cc0b0027-b82a-47a8-900c-6de076094417\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.921699 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-config\") pod \"cc0b0027-b82a-47a8-900c-6de076094417\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.921790 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-svc\") pod \"cc0b0027-b82a-47a8-900c-6de076094417\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.921806 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-sb\") pod \"cc0b0027-b82a-47a8-900c-6de076094417\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.921840 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-openstack-edpm-ipam\") pod \"cc0b0027-b82a-47a8-900c-6de076094417\" (UID: \"cc0b0027-b82a-47a8-900c-6de076094417\") " Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.929169 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc0b0027-b82a-47a8-900c-6de076094417-kube-api-access-b7vgm" (OuterVolumeSpecName: "kube-api-access-b7vgm") pod "cc0b0027-b82a-47a8-900c-6de076094417" (UID: "cc0b0027-b82a-47a8-900c-6de076094417"). InnerVolumeSpecName "kube-api-access-b7vgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.974642 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc0b0027-b82a-47a8-900c-6de076094417" (UID: "cc0b0027-b82a-47a8-900c-6de076094417"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.977758 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "cc0b0027-b82a-47a8-900c-6de076094417" (UID: "cc0b0027-b82a-47a8-900c-6de076094417"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.979008 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cc0b0027-b82a-47a8-900c-6de076094417" (UID: "cc0b0027-b82a-47a8-900c-6de076094417"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.980567 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cc0b0027-b82a-47a8-900c-6de076094417" (UID: "cc0b0027-b82a-47a8-900c-6de076094417"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.981915 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-config" (OuterVolumeSpecName: "config") pod "cc0b0027-b82a-47a8-900c-6de076094417" (UID: "cc0b0027-b82a-47a8-900c-6de076094417"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:35 crc kubenswrapper[4876]: I1205 08:40:35.984541 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cc0b0027-b82a-47a8-900c-6de076094417" (UID: "cc0b0027-b82a-47a8-900c-6de076094417"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.035269 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7vgm\" (UniqueName: \"kubernetes.io/projected/cc0b0027-b82a-47a8-900c-6de076094417-kube-api-access-b7vgm\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.035612 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.035627 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-config\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.035640 4876 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.035650 4876 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.035663 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.035675 4876 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc0b0027-b82a-47a8-900c-6de076094417-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.379258 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-g797c" event={"ID":"cc0b0027-b82a-47a8-900c-6de076094417","Type":"ContainerDied","Data":"05d39026a0ecc9a3f90678a38e9e846d262ae70fb009149b1c6d2e2fab0710b6"} Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.379314 4876 scope.go:117] "RemoveContainer" containerID="58b64f41cb01a3ac1c41bcfdc26018a761cfa3739bbb5910499c05cbc8a51689" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.379335 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-g797c" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.417464 4876 scope.go:117] "RemoveContainer" containerID="e9a6c2f76bea61e39105b3b1bafb425bb73d212aaf3a7ac0b6a0728afe29eb71" Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.425547 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-g797c"] Dec 05 08:40:36 crc kubenswrapper[4876]: I1205 08:40:36.456004 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-g797c"] Dec 05 08:40:37 crc kubenswrapper[4876]: I1205 08:40:37.833457 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc0b0027-b82a-47a8-900c-6de076094417" path="/var/lib/kubelet/pods/cc0b0027-b82a-47a8-900c-6de076094417/volumes" Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.213856 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.213937 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.213986 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.214763 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.214839 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" gracePeriod=600 Dec 05 08:40:38 crc kubenswrapper[4876]: E1205 08:40:38.386131 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.406155 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" exitCode=0 Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.406189 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8"} Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.406552 4876 scope.go:117] "RemoveContainer" containerID="9aa43e6caeca2c89f47c0cead848074edb893e13fe607a55570e6f144dca2a07" Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.407447 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:40:38 crc kubenswrapper[4876]: E1205 08:40:38.407797 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:40:38 crc kubenswrapper[4876]: I1205 08:40:38.992986 4876 scope.go:117] "RemoveContainer" containerID="a04f5489ec8e7b5f673febde7b0768b66fbed680c5ea48cbb16b83851551a9fa" Dec 05 08:40:39 crc kubenswrapper[4876]: I1205 08:40:39.029411 4876 scope.go:117] "RemoveContainer" containerID="99383be0e4efcabad0cf22226748cb90878ac40c5f56b578a56202004959f124" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.201456 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7tvpb"] Dec 05 08:40:41 crc kubenswrapper[4876]: E1205 08:40:41.202179 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" containerName="dnsmasq-dns" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.202197 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" containerName="dnsmasq-dns" Dec 05 08:40:41 crc kubenswrapper[4876]: E1205 08:40:41.202213 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" containerName="init" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.202220 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" containerName="init" Dec 05 08:40:41 crc kubenswrapper[4876]: E1205 08:40:41.202252 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0b0027-b82a-47a8-900c-6de076094417" containerName="dnsmasq-dns" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.202259 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0b0027-b82a-47a8-900c-6de076094417" containerName="dnsmasq-dns" Dec 05 08:40:41 crc kubenswrapper[4876]: E1205 08:40:41.202275 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0b0027-b82a-47a8-900c-6de076094417" containerName="init" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.202282 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0b0027-b82a-47a8-900c-6de076094417" containerName="init" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.202492 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0b0027-b82a-47a8-900c-6de076094417" containerName="dnsmasq-dns" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.202515 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db748c5-cab9-4494-8ba0-ff71f7b8b252" containerName="dnsmasq-dns" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.207453 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.222770 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tvpb"] Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.336845 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-catalog-content\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.336902 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-utilities\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.337084 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdbxl\" (UniqueName: \"kubernetes.io/projected/0e858930-ad7c-45a4-ad1d-903c666c7a80-kube-api-access-fdbxl\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.439353 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-catalog-content\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.439413 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-utilities\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.439566 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdbxl\" (UniqueName: \"kubernetes.io/projected/0e858930-ad7c-45a4-ad1d-903c666c7a80-kube-api-access-fdbxl\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.439950 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-catalog-content\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.439987 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-utilities\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.462243 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdbxl\" (UniqueName: \"kubernetes.io/projected/0e858930-ad7c-45a4-ad1d-903c666c7a80-kube-api-access-fdbxl\") pod \"redhat-marketplace-7tvpb\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.540826 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:41 crc kubenswrapper[4876]: I1205 08:40:41.994668 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tvpb"] Dec 05 08:40:42 crc kubenswrapper[4876]: W1205 08:40:42.001884 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e858930_ad7c_45a4_ad1d_903c666c7a80.slice/crio-55d2965a0c174fc641f369f92a0df0b6700cf3e0645492d1f5650170606c3a17 WatchSource:0}: Error finding container 55d2965a0c174fc641f369f92a0df0b6700cf3e0645492d1f5650170606c3a17: Status 404 returned error can't find the container with id 55d2965a0c174fc641f369f92a0df0b6700cf3e0645492d1f5650170606c3a17 Dec 05 08:40:42 crc kubenswrapper[4876]: I1205 08:40:42.447178 4876 generic.go:334] "Generic (PLEG): container finished" podID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerID="632d7ff5d7f09da4b1d4b51c49b00c7926bbccfdf0238be6ff8a1e42171ab090" exitCode=0 Dec 05 08:40:42 crc kubenswrapper[4876]: I1205 08:40:42.447268 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tvpb" event={"ID":"0e858930-ad7c-45a4-ad1d-903c666c7a80","Type":"ContainerDied","Data":"632d7ff5d7f09da4b1d4b51c49b00c7926bbccfdf0238be6ff8a1e42171ab090"} Dec 05 08:40:42 crc kubenswrapper[4876]: I1205 08:40:42.450035 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tvpb" event={"ID":"0e858930-ad7c-45a4-ad1d-903c666c7a80","Type":"ContainerStarted","Data":"55d2965a0c174fc641f369f92a0df0b6700cf3e0645492d1f5650170606c3a17"} Dec 05 08:40:43 crc kubenswrapper[4876]: I1205 08:40:43.462335 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tvpb" event={"ID":"0e858930-ad7c-45a4-ad1d-903c666c7a80","Type":"ContainerStarted","Data":"a6a572b55a22ab9712011c824b3cb0144e1c6c47d5ce3b83051a4d52a9acf112"} Dec 05 08:40:44 crc kubenswrapper[4876]: I1205 08:40:44.473803 4876 generic.go:334] "Generic (PLEG): container finished" podID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerID="a6a572b55a22ab9712011c824b3cb0144e1c6c47d5ce3b83051a4d52a9acf112" exitCode=0 Dec 05 08:40:44 crc kubenswrapper[4876]: I1205 08:40:44.473939 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tvpb" event={"ID":"0e858930-ad7c-45a4-ad1d-903c666c7a80","Type":"ContainerDied","Data":"a6a572b55a22ab9712011c824b3cb0144e1c6c47d5ce3b83051a4d52a9acf112"} Dec 05 08:40:45 crc kubenswrapper[4876]: I1205 08:40:45.498584 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tvpb" event={"ID":"0e858930-ad7c-45a4-ad1d-903c666c7a80","Type":"ContainerStarted","Data":"9c340bfd52fc60649aa29427949cd15f2024090853cd8cc70194bc798ec67679"} Dec 05 08:40:45 crc kubenswrapper[4876]: I1205 08:40:45.502261 4876 generic.go:334] "Generic (PLEG): container finished" podID="692469d9-b150-4d67-b412-ba8760598f8b" containerID="a6ff9c681250d1426f57c1ba83382bd24bc687513b76ea7c2e3db6bc5c50f907" exitCode=0 Dec 05 08:40:45 crc kubenswrapper[4876]: I1205 08:40:45.502302 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"692469d9-b150-4d67-b412-ba8760598f8b","Type":"ContainerDied","Data":"a6ff9c681250d1426f57c1ba83382bd24bc687513b76ea7c2e3db6bc5c50f907"} Dec 05 08:40:46 crc kubenswrapper[4876]: I1205 08:40:46.511646 4876 generic.go:334] "Generic (PLEG): container finished" podID="6ef895f6-3d13-4980-98ef-9111438fe1ec" containerID="2a3abf57515fd9df86edcaf44fe6c62332e0b96fc4c3e93ec0c054e366e7b26f" exitCode=0 Dec 05 08:40:46 crc kubenswrapper[4876]: I1205 08:40:46.511828 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6ef895f6-3d13-4980-98ef-9111438fe1ec","Type":"ContainerDied","Data":"2a3abf57515fd9df86edcaf44fe6c62332e0b96fc4c3e93ec0c054e366e7b26f"} Dec 05 08:40:46 crc kubenswrapper[4876]: I1205 08:40:46.572229 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7tvpb" podStartSLOduration=3.132210711 podStartE2EDuration="5.572206747s" podCreationTimestamp="2025-12-05 08:40:41 +0000 UTC" firstStartedPulling="2025-12-05 08:40:42.449561033 +0000 UTC m=+1506.938225695" lastFinishedPulling="2025-12-05 08:40:44.889557109 +0000 UTC m=+1509.378221731" observedRunningTime="2025-12-05 08:40:46.554430812 +0000 UTC m=+1511.043095434" watchObservedRunningTime="2025-12-05 08:40:46.572206747 +0000 UTC m=+1511.060871369" Dec 05 08:40:47 crc kubenswrapper[4876]: I1205 08:40:47.523498 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"692469d9-b150-4d67-b412-ba8760598f8b","Type":"ContainerStarted","Data":"1f60d159f8310274685ac179c0634ac90060b07cc1c477da0d0f4dc43b2fd061"} Dec 05 08:40:47 crc kubenswrapper[4876]: I1205 08:40:47.524080 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 08:40:47 crc kubenswrapper[4876]: I1205 08:40:47.525812 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6ef895f6-3d13-4980-98ef-9111438fe1ec","Type":"ContainerStarted","Data":"15e664310727c7d39a29a09c260fb11e2e215c261226384507b969eb99511ee3"} Dec 05 08:40:47 crc kubenswrapper[4876]: I1205 08:40:47.526108 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:40:47 crc kubenswrapper[4876]: I1205 08:40:47.557560 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.557536186 podStartE2EDuration="37.557536186s" podCreationTimestamp="2025-12-05 08:40:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:40:47.547042822 +0000 UTC m=+1512.035707474" watchObservedRunningTime="2025-12-05 08:40:47.557536186 +0000 UTC m=+1512.046200798" Dec 05 08:40:47 crc kubenswrapper[4876]: I1205 08:40:47.577009 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.576989844 podStartE2EDuration="37.576989844s" podCreationTimestamp="2025-12-05 08:40:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 08:40:47.574056318 +0000 UTC m=+1512.062720940" watchObservedRunningTime="2025-12-05 08:40:47.576989844 +0000 UTC m=+1512.065654476" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.592067 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz"] Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.593299 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.596253 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.597624 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.597710 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.597710 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.605061 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz"] Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.707202 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.707315 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.707376 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z26fz\" (UniqueName: \"kubernetes.io/projected/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-kube-api-access-z26fz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.707454 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.809495 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.809589 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.809661 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z26fz\" (UniqueName: \"kubernetes.io/projected/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-kube-api-access-z26fz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.809724 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.814728 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.815297 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.816445 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.832072 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z26fz\" (UniqueName: \"kubernetes.io/projected/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-kube-api-access-z26fz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:48 crc kubenswrapper[4876]: I1205 08:40:48.916948 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:40:49 crc kubenswrapper[4876]: I1205 08:40:49.629197 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz"] Dec 05 08:40:49 crc kubenswrapper[4876]: W1205 08:40:49.634485 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4554e5a3_3a60_4bc6_85c7_bfdc19421f05.slice/crio-ff14e86858d9bac9a777616b1c8302265bc84d378dde5a2554c1d8686c137591 WatchSource:0}: Error finding container ff14e86858d9bac9a777616b1c8302265bc84d378dde5a2554c1d8686c137591: Status 404 returned error can't find the container with id ff14e86858d9bac9a777616b1c8302265bc84d378dde5a2554c1d8686c137591 Dec 05 08:40:50 crc kubenswrapper[4876]: I1205 08:40:50.557204 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" event={"ID":"4554e5a3-3a60-4bc6-85c7-bfdc19421f05","Type":"ContainerStarted","Data":"ff14e86858d9bac9a777616b1c8302265bc84d378dde5a2554c1d8686c137591"} Dec 05 08:40:50 crc kubenswrapper[4876]: I1205 08:40:50.824517 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:40:50 crc kubenswrapper[4876]: E1205 08:40:50.824819 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:40:51 crc kubenswrapper[4876]: I1205 08:40:51.541816 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:51 crc kubenswrapper[4876]: I1205 08:40:51.542125 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:51 crc kubenswrapper[4876]: I1205 08:40:51.599185 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:51 crc kubenswrapper[4876]: I1205 08:40:51.657870 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:51 crc kubenswrapper[4876]: I1205 08:40:51.852682 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tvpb"] Dec 05 08:40:53 crc kubenswrapper[4876]: I1205 08:40:53.586786 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7tvpb" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerName="registry-server" containerID="cri-o://9c340bfd52fc60649aa29427949cd15f2024090853cd8cc70194bc798ec67679" gracePeriod=2 Dec 05 08:40:54 crc kubenswrapper[4876]: I1205 08:40:54.598038 4876 generic.go:334] "Generic (PLEG): container finished" podID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerID="9c340bfd52fc60649aa29427949cd15f2024090853cd8cc70194bc798ec67679" exitCode=0 Dec 05 08:40:54 crc kubenswrapper[4876]: I1205 08:40:54.598099 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tvpb" event={"ID":"0e858930-ad7c-45a4-ad1d-903c666c7a80","Type":"ContainerDied","Data":"9c340bfd52fc60649aa29427949cd15f2024090853cd8cc70194bc798ec67679"} Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.338532 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.423569 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-catalog-content\") pod \"0e858930-ad7c-45a4-ad1d-903c666c7a80\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.424021 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdbxl\" (UniqueName: \"kubernetes.io/projected/0e858930-ad7c-45a4-ad1d-903c666c7a80-kube-api-access-fdbxl\") pod \"0e858930-ad7c-45a4-ad1d-903c666c7a80\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.424185 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-utilities\") pod \"0e858930-ad7c-45a4-ad1d-903c666c7a80\" (UID: \"0e858930-ad7c-45a4-ad1d-903c666c7a80\") " Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.425165 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-utilities" (OuterVolumeSpecName: "utilities") pod "0e858930-ad7c-45a4-ad1d-903c666c7a80" (UID: "0e858930-ad7c-45a4-ad1d-903c666c7a80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.430144 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e858930-ad7c-45a4-ad1d-903c666c7a80-kube-api-access-fdbxl" (OuterVolumeSpecName: "kube-api-access-fdbxl") pod "0e858930-ad7c-45a4-ad1d-903c666c7a80" (UID: "0e858930-ad7c-45a4-ad1d-903c666c7a80"). InnerVolumeSpecName "kube-api-access-fdbxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.441300 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e858930-ad7c-45a4-ad1d-903c666c7a80" (UID: "0e858930-ad7c-45a4-ad1d-903c666c7a80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.527146 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdbxl\" (UniqueName: \"kubernetes.io/projected/0e858930-ad7c-45a4-ad1d-903c666c7a80-kube-api-access-fdbxl\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.527195 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.527209 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e858930-ad7c-45a4-ad1d-903c666c7a80-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.646430 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tvpb" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.646799 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tvpb" event={"ID":"0e858930-ad7c-45a4-ad1d-903c666c7a80","Type":"ContainerDied","Data":"55d2965a0c174fc641f369f92a0df0b6700cf3e0645492d1f5650170606c3a17"} Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.646868 4876 scope.go:117] "RemoveContainer" containerID="9c340bfd52fc60649aa29427949cd15f2024090853cd8cc70194bc798ec67679" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.648723 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" event={"ID":"4554e5a3-3a60-4bc6-85c7-bfdc19421f05","Type":"ContainerStarted","Data":"212b4d87e38d965a00e546fbeb53e1237d9a8657bc5d5ebf1e36d0311ee2a405"} Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.672543 4876 scope.go:117] "RemoveContainer" containerID="a6a572b55a22ab9712011c824b3cb0144e1c6c47d5ce3b83051a4d52a9acf112" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.693308 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" podStartSLOduration=2.293534481 podStartE2EDuration="11.693284473s" podCreationTimestamp="2025-12-05 08:40:48 +0000 UTC" firstStartedPulling="2025-12-05 08:40:49.637356626 +0000 UTC m=+1514.126021248" lastFinishedPulling="2025-12-05 08:40:59.037106618 +0000 UTC m=+1523.525771240" observedRunningTime="2025-12-05 08:40:59.672929721 +0000 UTC m=+1524.161594343" watchObservedRunningTime="2025-12-05 08:40:59.693284473 +0000 UTC m=+1524.181949095" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.698116 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tvpb"] Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.708579 4876 scope.go:117] "RemoveContainer" containerID="632d7ff5d7f09da4b1d4b51c49b00c7926bbccfdf0238be6ff8a1e42171ab090" Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.710731 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tvpb"] Dec 05 08:40:59 crc kubenswrapper[4876]: I1205 08:40:59.840291 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" path="/var/lib/kubelet/pods/0e858930-ad7c-45a4-ad1d-903c666c7a80/volumes" Dec 05 08:41:00 crc kubenswrapper[4876]: I1205 08:41:00.392169 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 08:41:00 crc kubenswrapper[4876]: I1205 08:41:00.780851 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 08:41:04 crc kubenswrapper[4876]: I1205 08:41:04.824838 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:41:04 crc kubenswrapper[4876]: E1205 08:41:04.825736 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:41:12 crc kubenswrapper[4876]: I1205 08:41:12.773104 4876 generic.go:334] "Generic (PLEG): container finished" podID="4554e5a3-3a60-4bc6-85c7-bfdc19421f05" containerID="212b4d87e38d965a00e546fbeb53e1237d9a8657bc5d5ebf1e36d0311ee2a405" exitCode=0 Dec 05 08:41:12 crc kubenswrapper[4876]: I1205 08:41:12.773188 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" event={"ID":"4554e5a3-3a60-4bc6-85c7-bfdc19421f05","Type":"ContainerDied","Data":"212b4d87e38d965a00e546fbeb53e1237d9a8657bc5d5ebf1e36d0311ee2a405"} Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.229763 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.336010 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory\") pod \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.336365 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z26fz\" (UniqueName: \"kubernetes.io/projected/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-kube-api-access-z26fz\") pod \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.336725 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-repo-setup-combined-ca-bundle\") pod \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.336830 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-ssh-key\") pod \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.350881 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-kube-api-access-z26fz" (OuterVolumeSpecName: "kube-api-access-z26fz") pod "4554e5a3-3a60-4bc6-85c7-bfdc19421f05" (UID: "4554e5a3-3a60-4bc6-85c7-bfdc19421f05"). InnerVolumeSpecName "kube-api-access-z26fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.351056 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4554e5a3-3a60-4bc6-85c7-bfdc19421f05" (UID: "4554e5a3-3a60-4bc6-85c7-bfdc19421f05"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:41:14 crc kubenswrapper[4876]: E1205 08:41:14.392399 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory podName:4554e5a3-3a60-4bc6-85c7-bfdc19421f05 nodeName:}" failed. No retries permitted until 2025-12-05 08:41:14.892362844 +0000 UTC m=+1539.381027486 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory") pod "4554e5a3-3a60-4bc6-85c7-bfdc19421f05" (UID: "4554e5a3-3a60-4bc6-85c7-bfdc19421f05") : error deleting /var/lib/kubelet/pods/4554e5a3-3a60-4bc6-85c7-bfdc19421f05/volume-subpaths: remove /var/lib/kubelet/pods/4554e5a3-3a60-4bc6-85c7-bfdc19421f05/volume-subpaths: no such file or directory Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.396654 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4554e5a3-3a60-4bc6-85c7-bfdc19421f05" (UID: "4554e5a3-3a60-4bc6-85c7-bfdc19421f05"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.439972 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.440033 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z26fz\" (UniqueName: \"kubernetes.io/projected/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-kube-api-access-z26fz\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.440049 4876 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.802303 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" event={"ID":"4554e5a3-3a60-4bc6-85c7-bfdc19421f05","Type":"ContainerDied","Data":"ff14e86858d9bac9a777616b1c8302265bc84d378dde5a2554c1d8686c137591"} Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.802660 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff14e86858d9bac9a777616b1c8302265bc84d378dde5a2554c1d8686c137591" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.802967 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.917038 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh"] Dec 05 08:41:14 crc kubenswrapper[4876]: E1205 08:41:14.917615 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerName="registry-server" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.917686 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerName="registry-server" Dec 05 08:41:14 crc kubenswrapper[4876]: E1205 08:41:14.917779 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4554e5a3-3a60-4bc6-85c7-bfdc19421f05" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.917838 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4554e5a3-3a60-4bc6-85c7-bfdc19421f05" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 08:41:14 crc kubenswrapper[4876]: E1205 08:41:14.917959 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerName="extract-utilities" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.918019 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerName="extract-utilities" Dec 05 08:41:14 crc kubenswrapper[4876]: E1205 08:41:14.918103 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerName="extract-content" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.918167 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerName="extract-content" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.918441 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e858930-ad7c-45a4-ad1d-903c666c7a80" containerName="registry-server" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.918535 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4554e5a3-3a60-4bc6-85c7-bfdc19421f05" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.919388 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.932014 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh"] Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.949086 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory\") pod \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\" (UID: \"4554e5a3-3a60-4bc6-85c7-bfdc19421f05\") " Dec 05 08:41:14 crc kubenswrapper[4876]: I1205 08:41:14.973213 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory" (OuterVolumeSpecName: "inventory") pod "4554e5a3-3a60-4bc6-85c7-bfdc19421f05" (UID: "4554e5a3-3a60-4bc6-85c7-bfdc19421f05"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.051838 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkgrr\" (UniqueName: \"kubernetes.io/projected/fde6a633-2ee9-4fad-b157-24435a2e89d4-kube-api-access-qkgrr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.052156 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.052339 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.052503 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4554e5a3-3a60-4bc6-85c7-bfdc19421f05-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.154659 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.155091 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkgrr\" (UniqueName: \"kubernetes.io/projected/fde6a633-2ee9-4fad-b157-24435a2e89d4-kube-api-access-qkgrr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.155232 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.159457 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.159988 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.178531 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkgrr\" (UniqueName: \"kubernetes.io/projected/fde6a633-2ee9-4fad-b157-24435a2e89d4-kube-api-access-qkgrr\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-7xhsh\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.330103 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:15 crc kubenswrapper[4876]: W1205 08:41:15.869264 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfde6a633_2ee9_4fad_b157_24435a2e89d4.slice/crio-eb18bfdb9c78aefe1227e0dea15c7ce513d418412bae7be99b3a42d50a6707c8 WatchSource:0}: Error finding container eb18bfdb9c78aefe1227e0dea15c7ce513d418412bae7be99b3a42d50a6707c8: Status 404 returned error can't find the container with id eb18bfdb9c78aefe1227e0dea15c7ce513d418412bae7be99b3a42d50a6707c8 Dec 05 08:41:15 crc kubenswrapper[4876]: I1205 08:41:15.872462 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh"] Dec 05 08:41:16 crc kubenswrapper[4876]: I1205 08:41:16.825749 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" event={"ID":"fde6a633-2ee9-4fad-b157-24435a2e89d4","Type":"ContainerStarted","Data":"4ee07f11e66f3d2cf20ece0f42636accd5a363e88fa36a60b0c8a098844fda43"} Dec 05 08:41:16 crc kubenswrapper[4876]: I1205 08:41:16.826353 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" event={"ID":"fde6a633-2ee9-4fad-b157-24435a2e89d4","Type":"ContainerStarted","Data":"eb18bfdb9c78aefe1227e0dea15c7ce513d418412bae7be99b3a42d50a6707c8"} Dec 05 08:41:16 crc kubenswrapper[4876]: I1205 08:41:16.864196 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" podStartSLOduration=2.457035707 podStartE2EDuration="2.864169658s" podCreationTimestamp="2025-12-05 08:41:14 +0000 UTC" firstStartedPulling="2025-12-05 08:41:15.870727247 +0000 UTC m=+1540.359391859" lastFinishedPulling="2025-12-05 08:41:16.277861188 +0000 UTC m=+1540.766525810" observedRunningTime="2025-12-05 08:41:16.845215264 +0000 UTC m=+1541.333879896" watchObservedRunningTime="2025-12-05 08:41:16.864169658 +0000 UTC m=+1541.352834291" Dec 05 08:41:18 crc kubenswrapper[4876]: I1205 08:41:18.823679 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:41:18 crc kubenswrapper[4876]: E1205 08:41:18.824251 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:41:19 crc kubenswrapper[4876]: I1205 08:41:19.861396 4876 generic.go:334] "Generic (PLEG): container finished" podID="fde6a633-2ee9-4fad-b157-24435a2e89d4" containerID="4ee07f11e66f3d2cf20ece0f42636accd5a363e88fa36a60b0c8a098844fda43" exitCode=0 Dec 05 08:41:19 crc kubenswrapper[4876]: I1205 08:41:19.861553 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" event={"ID":"fde6a633-2ee9-4fad-b157-24435a2e89d4","Type":"ContainerDied","Data":"4ee07f11e66f3d2cf20ece0f42636accd5a363e88fa36a60b0c8a098844fda43"} Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.274499 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.377764 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-inventory\") pod \"fde6a633-2ee9-4fad-b157-24435a2e89d4\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.377923 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-ssh-key\") pod \"fde6a633-2ee9-4fad-b157-24435a2e89d4\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.377983 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkgrr\" (UniqueName: \"kubernetes.io/projected/fde6a633-2ee9-4fad-b157-24435a2e89d4-kube-api-access-qkgrr\") pod \"fde6a633-2ee9-4fad-b157-24435a2e89d4\" (UID: \"fde6a633-2ee9-4fad-b157-24435a2e89d4\") " Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.391285 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde6a633-2ee9-4fad-b157-24435a2e89d4-kube-api-access-qkgrr" (OuterVolumeSpecName: "kube-api-access-qkgrr") pod "fde6a633-2ee9-4fad-b157-24435a2e89d4" (UID: "fde6a633-2ee9-4fad-b157-24435a2e89d4"). InnerVolumeSpecName "kube-api-access-qkgrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.407147 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-inventory" (OuterVolumeSpecName: "inventory") pod "fde6a633-2ee9-4fad-b157-24435a2e89d4" (UID: "fde6a633-2ee9-4fad-b157-24435a2e89d4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.407662 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fde6a633-2ee9-4fad-b157-24435a2e89d4" (UID: "fde6a633-2ee9-4fad-b157-24435a2e89d4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.479791 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.479825 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkgrr\" (UniqueName: \"kubernetes.io/projected/fde6a633-2ee9-4fad-b157-24435a2e89d4-kube-api-access-qkgrr\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.479836 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fde6a633-2ee9-4fad-b157-24435a2e89d4-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.882075 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" event={"ID":"fde6a633-2ee9-4fad-b157-24435a2e89d4","Type":"ContainerDied","Data":"eb18bfdb9c78aefe1227e0dea15c7ce513d418412bae7be99b3a42d50a6707c8"} Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.882117 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb18bfdb9c78aefe1227e0dea15c7ce513d418412bae7be99b3a42d50a6707c8" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.882436 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-7xhsh" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.954669 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj"] Dec 05 08:41:21 crc kubenswrapper[4876]: E1205 08:41:21.955256 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde6a633-2ee9-4fad-b157-24435a2e89d4" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.955352 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde6a633-2ee9-4fad-b157-24435a2e89d4" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.961619 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde6a633-2ee9-4fad-b157-24435a2e89d4" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.962915 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.966551 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.967240 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.967410 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.976355 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:41:21 crc kubenswrapper[4876]: I1205 08:41:21.993399 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj"] Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.088656 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.088712 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgh9v\" (UniqueName: \"kubernetes.io/projected/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-kube-api-access-qgh9v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.088753 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.088881 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.190695 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.190839 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.190869 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgh9v\" (UniqueName: \"kubernetes.io/projected/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-kube-api-access-qgh9v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.190916 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.194506 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.196094 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.198360 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.213990 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgh9v\" (UniqueName: \"kubernetes.io/projected/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-kube-api-access-qgh9v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.286996 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.817541 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj"] Dec 05 08:41:22 crc kubenswrapper[4876]: I1205 08:41:22.894593 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" event={"ID":"5f7a4eab-9eaf-4599-8eb6-deef38e971c9","Type":"ContainerStarted","Data":"d5442a204f9d82997a2abc0b7b58d6790bf83af47fe3e84bac9e3562da2ecee6"} Dec 05 08:41:25 crc kubenswrapper[4876]: I1205 08:41:25.935510 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" event={"ID":"5f7a4eab-9eaf-4599-8eb6-deef38e971c9","Type":"ContainerStarted","Data":"39b5a4d42e1a59d725f7cc546e8cf90ae5b9498ad3e1f61ca81be09745e34ed1"} Dec 05 08:41:25 crc kubenswrapper[4876]: I1205 08:41:25.957572 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" podStartSLOduration=3.009231473 podStartE2EDuration="4.95754808s" podCreationTimestamp="2025-12-05 08:41:21 +0000 UTC" firstStartedPulling="2025-12-05 08:41:22.828445281 +0000 UTC m=+1547.317109903" lastFinishedPulling="2025-12-05 08:41:24.776761888 +0000 UTC m=+1549.265426510" observedRunningTime="2025-12-05 08:41:25.950324502 +0000 UTC m=+1550.438989124" watchObservedRunningTime="2025-12-05 08:41:25.95754808 +0000 UTC m=+1550.446212702" Dec 05 08:41:27 crc kubenswrapper[4876]: I1205 08:41:27.840257 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jlxfs"] Dec 05 08:41:27 crc kubenswrapper[4876]: I1205 08:41:27.844353 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:27 crc kubenswrapper[4876]: I1205 08:41:27.846771 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jlxfs"] Dec 05 08:41:27 crc kubenswrapper[4876]: I1205 08:41:27.966086 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-utilities\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:27 crc kubenswrapper[4876]: I1205 08:41:27.966463 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-catalog-content\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:27 crc kubenswrapper[4876]: I1205 08:41:27.966601 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6rc5\" (UniqueName: \"kubernetes.io/projected/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-kube-api-access-r6rc5\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.068086 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6rc5\" (UniqueName: \"kubernetes.io/projected/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-kube-api-access-r6rc5\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.068189 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-utilities\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.068222 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-catalog-content\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.068687 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-catalog-content\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.068794 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-utilities\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.091180 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6rc5\" (UniqueName: \"kubernetes.io/projected/f48b9ece-bd4d-4f71-8238-99c5fdc0727a-kube-api-access-r6rc5\") pod \"certified-operators-jlxfs\" (UID: \"f48b9ece-bd4d-4f71-8238-99c5fdc0727a\") " pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.174310 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.728406 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jlxfs"] Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.967890 4876 generic.go:334] "Generic (PLEG): container finished" podID="f48b9ece-bd4d-4f71-8238-99c5fdc0727a" containerID="c2ed9ce2922b41125cef931a5b9934c74f89047123ef583780ed89eb5a83d526" exitCode=0 Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.968028 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlxfs" event={"ID":"f48b9ece-bd4d-4f71-8238-99c5fdc0727a","Type":"ContainerDied","Data":"c2ed9ce2922b41125cef931a5b9934c74f89047123ef583780ed89eb5a83d526"} Dec 05 08:41:28 crc kubenswrapper[4876]: I1205 08:41:28.968259 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlxfs" event={"ID":"f48b9ece-bd4d-4f71-8238-99c5fdc0727a","Type":"ContainerStarted","Data":"1e83e67c29bbf0e829ab0477aa1821350627108884d210d6be6568e7d44dc95b"} Dec 05 08:41:33 crc kubenswrapper[4876]: I1205 08:41:33.825720 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:41:33 crc kubenswrapper[4876]: E1205 08:41:33.826507 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:41:38 crc kubenswrapper[4876]: I1205 08:41:38.062346 4876 generic.go:334] "Generic (PLEG): container finished" podID="f48b9ece-bd4d-4f71-8238-99c5fdc0727a" containerID="75757d8c78db53c9713febaa7a04ff0ec05ef2a2eab6fd75f26c81010bae6e2b" exitCode=0 Dec 05 08:41:38 crc kubenswrapper[4876]: I1205 08:41:38.062414 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlxfs" event={"ID":"f48b9ece-bd4d-4f71-8238-99c5fdc0727a","Type":"ContainerDied","Data":"75757d8c78db53c9713febaa7a04ff0ec05ef2a2eab6fd75f26c81010bae6e2b"} Dec 05 08:41:39 crc kubenswrapper[4876]: I1205 08:41:39.077396 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jlxfs" event={"ID":"f48b9ece-bd4d-4f71-8238-99c5fdc0727a","Type":"ContainerStarted","Data":"01e5dcddb7d587f6e79acee3b281ffeaac52d52749d01b341f338b907ab5690a"} Dec 05 08:41:39 crc kubenswrapper[4876]: I1205 08:41:39.098428 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jlxfs" podStartSLOduration=2.436826681 podStartE2EDuration="12.098410812s" podCreationTimestamp="2025-12-05 08:41:27 +0000 UTC" firstStartedPulling="2025-12-05 08:41:28.969631634 +0000 UTC m=+1553.458296256" lastFinishedPulling="2025-12-05 08:41:38.631215765 +0000 UTC m=+1563.119880387" observedRunningTime="2025-12-05 08:41:39.096406798 +0000 UTC m=+1563.585071440" watchObservedRunningTime="2025-12-05 08:41:39.098410812 +0000 UTC m=+1563.587075434" Dec 05 08:41:39 crc kubenswrapper[4876]: I1205 08:41:39.196046 4876 scope.go:117] "RemoveContainer" containerID="baae878ad4ed798eea62443cd4ff0f9e29cbf336ba66ef9751d47a6ae8ade703" Dec 05 08:41:44 crc kubenswrapper[4876]: I1205 08:41:44.824305 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:41:44 crc kubenswrapper[4876]: E1205 08:41:44.825379 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:41:48 crc kubenswrapper[4876]: I1205 08:41:48.175152 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:48 crc kubenswrapper[4876]: I1205 08:41:48.176450 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:48 crc kubenswrapper[4876]: I1205 08:41:48.238370 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:49 crc kubenswrapper[4876]: I1205 08:41:49.213427 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jlxfs" Dec 05 08:41:49 crc kubenswrapper[4876]: I1205 08:41:49.374315 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jlxfs"] Dec 05 08:41:49 crc kubenswrapper[4876]: I1205 08:41:49.454626 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bl5c6"] Dec 05 08:41:49 crc kubenswrapper[4876]: I1205 08:41:49.455179 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bl5c6" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerName="registry-server" containerID="cri-o://7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6" gracePeriod=2 Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.085325 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.153748 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b2nh\" (UniqueName: \"kubernetes.io/projected/2c4bad1b-8683-4357-a021-cdd84678aa19-kube-api-access-4b2nh\") pod \"2c4bad1b-8683-4357-a021-cdd84678aa19\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.153879 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-catalog-content\") pod \"2c4bad1b-8683-4357-a021-cdd84678aa19\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.153945 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-utilities\") pod \"2c4bad1b-8683-4357-a021-cdd84678aa19\" (UID: \"2c4bad1b-8683-4357-a021-cdd84678aa19\") " Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.154823 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-utilities" (OuterVolumeSpecName: "utilities") pod "2c4bad1b-8683-4357-a021-cdd84678aa19" (UID: "2c4bad1b-8683-4357-a021-cdd84678aa19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.160370 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c4bad1b-8683-4357-a021-cdd84678aa19-kube-api-access-4b2nh" (OuterVolumeSpecName: "kube-api-access-4b2nh") pod "2c4bad1b-8683-4357-a021-cdd84678aa19" (UID: "2c4bad1b-8683-4357-a021-cdd84678aa19"). InnerVolumeSpecName "kube-api-access-4b2nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.179389 4876 generic.go:334] "Generic (PLEG): container finished" podID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerID="7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6" exitCode=0 Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.179463 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bl5c6" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.179505 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bl5c6" event={"ID":"2c4bad1b-8683-4357-a021-cdd84678aa19","Type":"ContainerDied","Data":"7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6"} Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.179533 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bl5c6" event={"ID":"2c4bad1b-8683-4357-a021-cdd84678aa19","Type":"ContainerDied","Data":"dd9b95c8f0be41a43cbb084d0b0f980c76ee3cfba90eafc3661e1dd224d1daf3"} Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.179550 4876 scope.go:117] "RemoveContainer" containerID="7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.211445 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c4bad1b-8683-4357-a021-cdd84678aa19" (UID: "2c4bad1b-8683-4357-a021-cdd84678aa19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.231533 4876 scope.go:117] "RemoveContainer" containerID="2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.256173 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.256406 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c4bad1b-8683-4357-a021-cdd84678aa19-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.256462 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b2nh\" (UniqueName: \"kubernetes.io/projected/2c4bad1b-8683-4357-a021-cdd84678aa19-kube-api-access-4b2nh\") on node \"crc\" DevicePath \"\"" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.257890 4876 scope.go:117] "RemoveContainer" containerID="8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.296881 4876 scope.go:117] "RemoveContainer" containerID="7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6" Dec 05 08:41:50 crc kubenswrapper[4876]: E1205 08:41:50.297410 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6\": container with ID starting with 7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6 not found: ID does not exist" containerID="7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.297507 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6"} err="failed to get container status \"7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6\": rpc error: code = NotFound desc = could not find container \"7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6\": container with ID starting with 7a1d6a8e1f69ba64e7176db4d184a58fc60158c294608c7f9ba052e68f70a3c6 not found: ID does not exist" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.297677 4876 scope.go:117] "RemoveContainer" containerID="2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c" Dec 05 08:41:50 crc kubenswrapper[4876]: E1205 08:41:50.298177 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c\": container with ID starting with 2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c not found: ID does not exist" containerID="2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.298259 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c"} err="failed to get container status \"2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c\": rpc error: code = NotFound desc = could not find container \"2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c\": container with ID starting with 2adab5bddd73893ec61f6a2f8f6ef10099a056ccbdc8796cf8e215579a74162c not found: ID does not exist" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.298331 4876 scope.go:117] "RemoveContainer" containerID="8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b" Dec 05 08:41:50 crc kubenswrapper[4876]: E1205 08:41:50.298700 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b\": container with ID starting with 8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b not found: ID does not exist" containerID="8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.298792 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b"} err="failed to get container status \"8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b\": rpc error: code = NotFound desc = could not find container \"8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b\": container with ID starting with 8e8ffb680b091e560293e975737c391269b54adffb23da87cedec7052a5a3b5b not found: ID does not exist" Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.521781 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bl5c6"] Dec 05 08:41:50 crc kubenswrapper[4876]: I1205 08:41:50.534552 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bl5c6"] Dec 05 08:41:51 crc kubenswrapper[4876]: I1205 08:41:51.840427 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" path="/var/lib/kubelet/pods/2c4bad1b-8683-4357-a021-cdd84678aa19/volumes" Dec 05 08:41:56 crc kubenswrapper[4876]: I1205 08:41:56.823668 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:41:56 crc kubenswrapper[4876]: E1205 08:41:56.824401 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:42:07 crc kubenswrapper[4876]: I1205 08:42:07.830054 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:42:07 crc kubenswrapper[4876]: E1205 08:42:07.830888 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:42:21 crc kubenswrapper[4876]: I1205 08:42:21.823989 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:42:21 crc kubenswrapper[4876]: E1205 08:42:21.824803 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:42:32 crc kubenswrapper[4876]: I1205 08:42:32.823873 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:42:32 crc kubenswrapper[4876]: E1205 08:42:32.825886 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:42:39 crc kubenswrapper[4876]: I1205 08:42:39.290188 4876 scope.go:117] "RemoveContainer" containerID="bfa61edc4a079881d538371e415544a2e93a9c3a589d05c7f54e6c19f38443ae" Dec 05 08:42:44 crc kubenswrapper[4876]: I1205 08:42:44.824812 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:42:44 crc kubenswrapper[4876]: E1205 08:42:44.825869 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:42:56 crc kubenswrapper[4876]: I1205 08:42:56.824014 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:42:56 crc kubenswrapper[4876]: E1205 08:42:56.824744 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:43:07 crc kubenswrapper[4876]: I1205 08:43:07.825992 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:43:07 crc kubenswrapper[4876]: E1205 08:43:07.829518 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:43:19 crc kubenswrapper[4876]: I1205 08:43:19.824754 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:43:19 crc kubenswrapper[4876]: E1205 08:43:19.825933 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:43:33 crc kubenswrapper[4876]: I1205 08:43:33.824301 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:43:33 crc kubenswrapper[4876]: E1205 08:43:33.825081 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:43:48 crc kubenswrapper[4876]: I1205 08:43:48.825756 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:43:48 crc kubenswrapper[4876]: E1205 08:43:48.827191 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:43:59 crc kubenswrapper[4876]: I1205 08:43:59.824820 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:43:59 crc kubenswrapper[4876]: E1205 08:43:59.825529 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:44:12 crc kubenswrapper[4876]: I1205 08:44:12.823708 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:44:12 crc kubenswrapper[4876]: E1205 08:44:12.824562 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:44:23 crc kubenswrapper[4876]: I1205 08:44:23.823943 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:44:23 crc kubenswrapper[4876]: E1205 08:44:23.826035 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:44:36 crc kubenswrapper[4876]: I1205 08:44:36.825223 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:44:36 crc kubenswrapper[4876]: E1205 08:44:36.826065 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:44:39 crc kubenswrapper[4876]: I1205 08:44:39.401826 4876 scope.go:117] "RemoveContainer" containerID="6d5ddaea1f3d3216e45035a5b36287b4b0b2809faf1e49b87cd4d9df114e622d" Dec 05 08:44:39 crc kubenswrapper[4876]: I1205 08:44:39.435032 4876 scope.go:117] "RemoveContainer" containerID="5a342e34f983f5092466e9864f0f8c3cf28865539e000d238e11d3aaf9f6ab15" Dec 05 08:44:46 crc kubenswrapper[4876]: I1205 08:44:46.002454 4876 generic.go:334] "Generic (PLEG): container finished" podID="5f7a4eab-9eaf-4599-8eb6-deef38e971c9" containerID="39b5a4d42e1a59d725f7cc546e8cf90ae5b9498ad3e1f61ca81be09745e34ed1" exitCode=0 Dec 05 08:44:46 crc kubenswrapper[4876]: I1205 08:44:46.002525 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" event={"ID":"5f7a4eab-9eaf-4599-8eb6-deef38e971c9","Type":"ContainerDied","Data":"39b5a4d42e1a59d725f7cc546e8cf90ae5b9498ad3e1f61ca81be09745e34ed1"} Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.455395 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.490819 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-ssh-key\") pod \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.490950 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-bootstrap-combined-ca-bundle\") pod \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.490976 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgh9v\" (UniqueName: \"kubernetes.io/projected/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-kube-api-access-qgh9v\") pod \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.491037 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-inventory\") pod \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\" (UID: \"5f7a4eab-9eaf-4599-8eb6-deef38e971c9\") " Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.497413 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-kube-api-access-qgh9v" (OuterVolumeSpecName: "kube-api-access-qgh9v") pod "5f7a4eab-9eaf-4599-8eb6-deef38e971c9" (UID: "5f7a4eab-9eaf-4599-8eb6-deef38e971c9"). InnerVolumeSpecName "kube-api-access-qgh9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.497445 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5f7a4eab-9eaf-4599-8eb6-deef38e971c9" (UID: "5f7a4eab-9eaf-4599-8eb6-deef38e971c9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.524057 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f7a4eab-9eaf-4599-8eb6-deef38e971c9" (UID: "5f7a4eab-9eaf-4599-8eb6-deef38e971c9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.525576 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-inventory" (OuterVolumeSpecName: "inventory") pod "5f7a4eab-9eaf-4599-8eb6-deef38e971c9" (UID: "5f7a4eab-9eaf-4599-8eb6-deef38e971c9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.592946 4876 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.592989 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgh9v\" (UniqueName: \"kubernetes.io/projected/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-kube-api-access-qgh9v\") on node \"crc\" DevicePath \"\"" Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.593006 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:44:47 crc kubenswrapper[4876]: I1205 08:44:47.593016 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f7a4eab-9eaf-4599-8eb6-deef38e971c9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.026356 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" event={"ID":"5f7a4eab-9eaf-4599-8eb6-deef38e971c9","Type":"ContainerDied","Data":"d5442a204f9d82997a2abc0b7b58d6790bf83af47fe3e84bac9e3562da2ecee6"} Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.026825 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5442a204f9d82997a2abc0b7b58d6790bf83af47fe3e84bac9e3562da2ecee6" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.026408 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.127757 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn"] Dec 05 08:44:48 crc kubenswrapper[4876]: E1205 08:44:48.128278 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerName="extract-content" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.128303 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerName="extract-content" Dec 05 08:44:48 crc kubenswrapper[4876]: E1205 08:44:48.128325 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerName="registry-server" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.128333 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerName="registry-server" Dec 05 08:44:48 crc kubenswrapper[4876]: E1205 08:44:48.128349 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7a4eab-9eaf-4599-8eb6-deef38e971c9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.128361 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7a4eab-9eaf-4599-8eb6-deef38e971c9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 08:44:48 crc kubenswrapper[4876]: E1205 08:44:48.128379 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerName="extract-utilities" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.128393 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerName="extract-utilities" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.128627 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c4bad1b-8683-4357-a021-cdd84678aa19" containerName="registry-server" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.128650 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7a4eab-9eaf-4599-8eb6-deef38e971c9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.129376 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.131189 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.131401 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.132372 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.142362 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.150835 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn"] Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.313254 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxdtd\" (UniqueName: \"kubernetes.io/projected/d17d4141-7f2f-4598-9cf4-96870bca5903-kube-api-access-cxdtd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.313459 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.313531 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.414746 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxdtd\" (UniqueName: \"kubernetes.io/projected/d17d4141-7f2f-4598-9cf4-96870bca5903-kube-api-access-cxdtd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.414846 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.414876 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.418869 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.431114 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.432835 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxdtd\" (UniqueName: \"kubernetes.io/projected/d17d4141-7f2f-4598-9cf4-96870bca5903-kube-api-access-cxdtd\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.445851 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.989764 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn"] Dec 05 08:44:48 crc kubenswrapper[4876]: I1205 08:44:48.992266 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:44:49 crc kubenswrapper[4876]: I1205 08:44:49.037199 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" event={"ID":"d17d4141-7f2f-4598-9cf4-96870bca5903","Type":"ContainerStarted","Data":"b1b9a7ebb2267d033db3994bdf49644e8c9f93a54890c0001be262f24c7e87f2"} Dec 05 08:44:49 crc kubenswrapper[4876]: I1205 08:44:49.824703 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:44:49 crc kubenswrapper[4876]: E1205 08:44:49.825556 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:44:50 crc kubenswrapper[4876]: I1205 08:44:50.047503 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" event={"ID":"d17d4141-7f2f-4598-9cf4-96870bca5903","Type":"ContainerStarted","Data":"ff425636f2433d407f59a4a5c3499d872f2835ab13daeff8512cc9db1fd80d10"} Dec 05 08:44:50 crc kubenswrapper[4876]: I1205 08:44:50.069615 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" podStartSLOduration=1.582489753 podStartE2EDuration="2.069596852s" podCreationTimestamp="2025-12-05 08:44:48 +0000 UTC" firstStartedPulling="2025-12-05 08:44:48.991869497 +0000 UTC m=+1753.480534129" lastFinishedPulling="2025-12-05 08:44:49.478976606 +0000 UTC m=+1753.967641228" observedRunningTime="2025-12-05 08:44:50.067940177 +0000 UTC m=+1754.556604829" watchObservedRunningTime="2025-12-05 08:44:50.069596852 +0000 UTC m=+1754.558261474" Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.048755 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bwggg"] Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.063050 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-w4xpx"] Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.073980 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-96e4-account-create-update-6fzmf"] Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.083177 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-348b-account-create-update-5dmk7"] Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.094045 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bwggg"] Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.102092 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-96e4-account-create-update-6fzmf"] Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.112134 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-w4xpx"] Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.122509 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-348b-account-create-update-5dmk7"] Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.835958 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee916a3-9c6b-4102-84b1-f921c334f152" path="/var/lib/kubelet/pods/2ee916a3-9c6b-4102-84b1-f921c334f152/volumes" Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.837148 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90c4575e-5eb6-4c24-8662-9c2c4863bc08" path="/var/lib/kubelet/pods/90c4575e-5eb6-4c24-8662-9c2c4863bc08/volumes" Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.837792 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8b76471-1845-41af-aecd-ea80bf5ba97d" path="/var/lib/kubelet/pods/b8b76471-1845-41af-aecd-ea80bf5ba97d/volumes" Dec 05 08:44:55 crc kubenswrapper[4876]: I1205 08:44:55.838621 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee" path="/var/lib/kubelet/pods/f2a03ebb-43ac-4d3e-a548-6ae7c48d22ee/volumes" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.048406 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-mq6ht"] Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.059105 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-7dd9-account-create-update-k5wq2"] Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.067760 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-mq6ht"] Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.075909 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-7dd9-account-create-update-k5wq2"] Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.140881 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq"] Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.142489 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.144342 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.144974 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.151530 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq"] Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.242168 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgklh\" (UniqueName: \"kubernetes.io/projected/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-kube-api-access-bgklh\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.242476 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-config-volume\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.242719 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-secret-volume\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.345261 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-config-volume\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.345450 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-secret-volume\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.345600 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgklh\" (UniqueName: \"kubernetes.io/projected/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-kube-api-access-bgklh\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.346696 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-config-volume\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.353622 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-secret-volume\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.365931 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgklh\" (UniqueName: \"kubernetes.io/projected/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-kube-api-access-bgklh\") pod \"collect-profiles-29415405-2hpzq\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.466665 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.824005 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:45:00 crc kubenswrapper[4876]: E1205 08:45:00.824715 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:45:00 crc kubenswrapper[4876]: I1205 08:45:00.946380 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq"] Dec 05 08:45:01 crc kubenswrapper[4876]: I1205 08:45:01.152684 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" event={"ID":"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c","Type":"ContainerStarted","Data":"d1782f0f06de456c949ae455429bbd98e9c267814cd57f9f7a5bdc239badfb1c"} Dec 05 08:45:01 crc kubenswrapper[4876]: I1205 08:45:01.153158 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" event={"ID":"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c","Type":"ContainerStarted","Data":"dbaac8618a2be6d03453e95fec91da5ec652dee543c3f6b037b5fe29f6b895bc"} Dec 05 08:45:01 crc kubenswrapper[4876]: I1205 08:45:01.835729 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d158569-ccc7-4670-836b-c3459be0f527" path="/var/lib/kubelet/pods/4d158569-ccc7-4670-836b-c3459be0f527/volumes" Dec 05 08:45:01 crc kubenswrapper[4876]: I1205 08:45:01.836674 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de5380be-9b7d-4320-ae64-a98efdafcb82" path="/var/lib/kubelet/pods/de5380be-9b7d-4320-ae64-a98efdafcb82/volumes" Dec 05 08:45:02 crc kubenswrapper[4876]: I1205 08:45:02.174644 4876 generic.go:334] "Generic (PLEG): container finished" podID="e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c" containerID="d1782f0f06de456c949ae455429bbd98e9c267814cd57f9f7a5bdc239badfb1c" exitCode=0 Dec 05 08:45:02 crc kubenswrapper[4876]: I1205 08:45:02.174692 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" event={"ID":"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c","Type":"ContainerDied","Data":"d1782f0f06de456c949ae455429bbd98e9c267814cd57f9f7a5bdc239badfb1c"} Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.558040 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.608258 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-config-volume\") pod \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.608331 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-secret-volume\") pod \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.608681 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgklh\" (UniqueName: \"kubernetes.io/projected/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-kube-api-access-bgklh\") pod \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\" (UID: \"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c\") " Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.609248 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-config-volume" (OuterVolumeSpecName: "config-volume") pod "e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c" (UID: "e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.609997 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.618683 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c" (UID: "e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.621709 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-kube-api-access-bgklh" (OuterVolumeSpecName: "kube-api-access-bgklh") pod "e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c" (UID: "e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c"). InnerVolumeSpecName "kube-api-access-bgklh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.710876 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgklh\" (UniqueName: \"kubernetes.io/projected/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-kube-api-access-bgklh\") on node \"crc\" DevicePath \"\"" Dec 05 08:45:03 crc kubenswrapper[4876]: I1205 08:45:03.711196 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 08:45:04 crc kubenswrapper[4876]: I1205 08:45:04.197821 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" event={"ID":"e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c","Type":"ContainerDied","Data":"dbaac8618a2be6d03453e95fec91da5ec652dee543c3f6b037b5fe29f6b895bc"} Dec 05 08:45:04 crc kubenswrapper[4876]: I1205 08:45:04.197867 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbaac8618a2be6d03453e95fec91da5ec652dee543c3f6b037b5fe29f6b895bc" Dec 05 08:45:04 crc kubenswrapper[4876]: I1205 08:45:04.197975 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq" Dec 05 08:45:13 crc kubenswrapper[4876]: I1205 08:45:13.824433 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:45:13 crc kubenswrapper[4876]: E1205 08:45:13.825245 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:45:24 crc kubenswrapper[4876]: I1205 08:45:24.824056 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:45:24 crc kubenswrapper[4876]: E1205 08:45:24.824980 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.072353 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-n7phf"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.090507 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-mm4br"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.098621 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9kk66"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.114121 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5909-account-create-update-6pw92"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.124024 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8029-account-create-update-j2rvm"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.133927 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6e8e-account-create-update-jnkdv"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.142137 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-xz6wd"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.151085 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5909-account-create-update-6pw92"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.159353 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-mm4br"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.168580 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-bzms5"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.180961 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8029-account-create-update-j2rvm"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.193106 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-bzms5"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.201410 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6e8e-account-create-update-jnkdv"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.209755 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-xz6wd"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.218011 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-a905-account-create-update-zr6j4"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.226325 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9kk66"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.235101 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-n7phf"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.242759 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-a905-account-create-update-zr6j4"] Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.850592 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be" path="/var/lib/kubelet/pods/08070b8a-b86a-4ab6-ab8e-6cc1cf64e4be/volumes" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.851358 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f" path="/var/lib/kubelet/pods/1ad4fc3c-ce8e-445b-8f44-1bf9a6c9874f/volumes" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.852183 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c58acae-5113-4e33-951e-409855923e87" path="/var/lib/kubelet/pods/2c58acae-5113-4e33-951e-409855923e87/volumes" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.852983 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="327b4198-11d6-4b61-829e-ef09616182b4" path="/var/lib/kubelet/pods/327b4198-11d6-4b61-829e-ef09616182b4/volumes" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.854424 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e46c2c-ff07-4431-ab74-fafdc1581673" path="/var/lib/kubelet/pods/80e46c2c-ff07-4431-ab74-fafdc1581673/volumes" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.855536 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9855cf16-5f53-4795-9103-3ee2c6b75f79" path="/var/lib/kubelet/pods/9855cf16-5f53-4795-9103-3ee2c6b75f79/volumes" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.856142 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6" path="/var/lib/kubelet/pods/bd933f46-dcb8-4ad4-9554-67c5d2a3d7c6/volumes" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.857659 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dabfca06-e3fb-4f63-baaf-f6b24bd4460a" path="/var/lib/kubelet/pods/dabfca06-e3fb-4f63-baaf-f6b24bd4460a/volumes" Dec 05 08:45:35 crc kubenswrapper[4876]: I1205 08:45:35.858258 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db2de57d-f90f-44c1-89de-d7706bd6e64c" path="/var/lib/kubelet/pods/db2de57d-f90f-44c1-89de-d7706bd6e64c/volumes" Dec 05 08:45:38 crc kubenswrapper[4876]: I1205 08:45:38.824329 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.513650 4876 scope.go:117] "RemoveContainer" containerID="ddbc9e75a03fa7ff0cec4ff0c2d5ee17066f3aa0201e614f12a1dc1f18d4cd61" Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.556047 4876 scope.go:117] "RemoveContainer" containerID="7905d76b31b558ffe5a19de47f7f4ff57d1727f04a59c68bae10b5b238ff26b2" Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.559339 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"5ca3218eccf9f7164d33de275ebcdaf5bcf4922b876f83c58de4aed55797c0a8"} Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.659573 4876 scope.go:117] "RemoveContainer" containerID="3eb6f5ab627df8b351bd9b5d1b6b6df495fc5d8862ae6dee1286d23f97eebf8f" Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.804165 4876 scope.go:117] "RemoveContainer" containerID="bfe4c276c30394787954befcccf806035dc8eef3b38dddf3dc8b4e446b5807c4" Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.877914 4876 scope.go:117] "RemoveContainer" containerID="7b3ddfe2c86d27790ba0d10e5040d810ae8e8bd436c5fd8db96f31dfbc715646" Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.904166 4876 scope.go:117] "RemoveContainer" containerID="773631e225f32fc1597c11685caa78682e0a9bee7833e77607fe9a4583a06da9" Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.952706 4876 scope.go:117] "RemoveContainer" containerID="2114de6fdf9833669bc77e6056c62bdef51b47d93782be620403cba1524f562b" Dec 05 08:45:39 crc kubenswrapper[4876]: I1205 08:45:39.994249 4876 scope.go:117] "RemoveContainer" containerID="7f5519a23a220ed25975e777bd47418dc954cbbbf9d4757dc699d7babaf9ac20" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.044126 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-rptl7"] Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.051403 4876 scope.go:117] "RemoveContainer" containerID="7b32f9d9dc611194a0a9ab986b1d4f84d0211cff6ec57beee41b83b486299032" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.055329 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-rptl7"] Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.127206 4876 scope.go:117] "RemoveContainer" containerID="70660ff27709f826aeef1b81f8f9df35fdcf1e65df3f8281a0a8ef2695ecfb58" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.164976 4876 scope.go:117] "RemoveContainer" containerID="d206ea9f116f0d29a176869b29933ae54243c8aaa83d42034b916c1d98b04679" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.188674 4876 scope.go:117] "RemoveContainer" containerID="57777bc0ed58a1d0d7d50bdd58c5c6f106344fcfa8a341227b49ee4cdab7ba55" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.216198 4876 scope.go:117] "RemoveContainer" containerID="6c3ff6b383912f24151094bdbc91f3b7716fc94c398919c79a9b60cf6e5a50ea" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.247474 4876 scope.go:117] "RemoveContainer" containerID="0c6280ae9e8c0648a1c1b8d7d9bb7e9444a1282df9ea731bfc4f64968c18b5ec" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.273624 4876 scope.go:117] "RemoveContainer" containerID="bcea7a3dc3026cc23d48072974b77b7de8dd3d3b624c940ed7f0d2cf2e64666f" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.296243 4876 scope.go:117] "RemoveContainer" containerID="f33eb3e638dd9f48068de5845629722c4abfbfb5789913f149f65ca34b79cb20" Dec 05 08:45:40 crc kubenswrapper[4876]: I1205 08:45:40.318475 4876 scope.go:117] "RemoveContainer" containerID="2f8e03026f257a88664deb69ca0e4180b7c2c4fd986841ab7e9c921ac63b1174" Dec 05 08:45:41 crc kubenswrapper[4876]: I1205 08:45:41.845278 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef68788-39b1-4011-bd01-a63d7520f540" path="/var/lib/kubelet/pods/5ef68788-39b1-4011-bd01-a63d7520f540/volumes" Dec 05 08:46:08 crc kubenswrapper[4876]: I1205 08:46:08.050467 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2rwzp"] Dec 05 08:46:08 crc kubenswrapper[4876]: I1205 08:46:08.068677 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2rwzp"] Dec 05 08:46:09 crc kubenswrapper[4876]: I1205 08:46:09.849067 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="207c5e9c-418a-4c07-8e8a-bb79948d855d" path="/var/lib/kubelet/pods/207c5e9c-418a-4c07-8e8a-bb79948d855d/volumes" Dec 05 08:46:17 crc kubenswrapper[4876]: I1205 08:46:17.028814 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wc9k9"] Dec 05 08:46:17 crc kubenswrapper[4876]: I1205 08:46:17.042929 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wc9k9"] Dec 05 08:46:17 crc kubenswrapper[4876]: I1205 08:46:17.836253 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1" path="/var/lib/kubelet/pods/9187c3bd-2fa8-41dd-a669-09d0a4fd7cd1/volumes" Dec 05 08:46:20 crc kubenswrapper[4876]: I1205 08:46:20.032631 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-n5bmx"] Dec 05 08:46:20 crc kubenswrapper[4876]: I1205 08:46:20.042238 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-m2khz"] Dec 05 08:46:20 crc kubenswrapper[4876]: I1205 08:46:20.051381 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-n5bmx"] Dec 05 08:46:20 crc kubenswrapper[4876]: I1205 08:46:20.060132 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-m2khz"] Dec 05 08:46:21 crc kubenswrapper[4876]: I1205 08:46:21.837632 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d18b829-e516-41de-989d-bea32b940409" path="/var/lib/kubelet/pods/2d18b829-e516-41de-989d-bea32b940409/volumes" Dec 05 08:46:21 crc kubenswrapper[4876]: I1205 08:46:21.838572 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="460f93bd-e732-42c0-8e75-625f3346800c" path="/var/lib/kubelet/pods/460f93bd-e732-42c0-8e75-625f3346800c/volumes" Dec 05 08:46:30 crc kubenswrapper[4876]: I1205 08:46:30.038582 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-xbjb2"] Dec 05 08:46:30 crc kubenswrapper[4876]: I1205 08:46:30.047983 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vxcln"] Dec 05 08:46:30 crc kubenswrapper[4876]: I1205 08:46:30.059086 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vxcln"] Dec 05 08:46:30 crc kubenswrapper[4876]: I1205 08:46:30.067529 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-xbjb2"] Dec 05 08:46:31 crc kubenswrapper[4876]: I1205 08:46:31.840327 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1ae752-b331-4d0f-8039-e5720312f674" path="/var/lib/kubelet/pods/5a1ae752-b331-4d0f-8039-e5720312f674/volumes" Dec 05 08:46:31 crc kubenswrapper[4876]: I1205 08:46:31.841952 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62f1d7bc-1a36-4fc0-a94d-72fcedda052e" path="/var/lib/kubelet/pods/62f1d7bc-1a36-4fc0-a94d-72fcedda052e/volumes" Dec 05 08:46:40 crc kubenswrapper[4876]: I1205 08:46:40.588666 4876 scope.go:117] "RemoveContainer" containerID="cbc720abbc55b77c1cbdf9bb522e82ac772b3b46fdfb8f291d1fb783bcaf5ad2" Dec 05 08:46:40 crc kubenswrapper[4876]: I1205 08:46:40.647256 4876 scope.go:117] "RemoveContainer" containerID="5b43864474344dca68358c926da69812978879965307a50d4093bbef607f1897" Dec 05 08:46:40 crc kubenswrapper[4876]: I1205 08:46:40.699385 4876 scope.go:117] "RemoveContainer" containerID="d0694e5b242890e48a9b0af47bb44693b50de3adca8d565e9f7a01d0e76ac914" Dec 05 08:46:40 crc kubenswrapper[4876]: I1205 08:46:40.743804 4876 scope.go:117] "RemoveContainer" containerID="ba7f7adc5c72b82dec3c9ece793de8e91289022fd9785b8cf8677e04284a946e" Dec 05 08:46:40 crc kubenswrapper[4876]: I1205 08:46:40.775478 4876 scope.go:117] "RemoveContainer" containerID="d5b7822ecdb59c6b8b44f0f4c15198949fd589345f3891c5cb7760da54614320" Dec 05 08:46:40 crc kubenswrapper[4876]: I1205 08:46:40.874111 4876 scope.go:117] "RemoveContainer" containerID="c0c041a52c79ac54e83bfddc4d7fea542eb2845f3ca1890eeedb1ebd3a60ab32" Dec 05 08:46:40 crc kubenswrapper[4876]: I1205 08:46:40.908670 4876 scope.go:117] "RemoveContainer" containerID="215e4e1b05c2f0f608af1bb1647a803db99e077dabcf6a941b8d8a1170b223fc" Dec 05 08:46:45 crc kubenswrapper[4876]: I1205 08:46:45.168580 4876 generic.go:334] "Generic (PLEG): container finished" podID="d17d4141-7f2f-4598-9cf4-96870bca5903" containerID="ff425636f2433d407f59a4a5c3499d872f2835ab13daeff8512cc9db1fd80d10" exitCode=0 Dec 05 08:46:45 crc kubenswrapper[4876]: I1205 08:46:45.168675 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" event={"ID":"d17d4141-7f2f-4598-9cf4-96870bca5903","Type":"ContainerDied","Data":"ff425636f2433d407f59a4a5c3499d872f2835ab13daeff8512cc9db1fd80d10"} Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.619215 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.790356 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-ssh-key\") pod \"d17d4141-7f2f-4598-9cf4-96870bca5903\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.790555 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-inventory\") pod \"d17d4141-7f2f-4598-9cf4-96870bca5903\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.790649 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxdtd\" (UniqueName: \"kubernetes.io/projected/d17d4141-7f2f-4598-9cf4-96870bca5903-kube-api-access-cxdtd\") pod \"d17d4141-7f2f-4598-9cf4-96870bca5903\" (UID: \"d17d4141-7f2f-4598-9cf4-96870bca5903\") " Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.807239 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17d4141-7f2f-4598-9cf4-96870bca5903-kube-api-access-cxdtd" (OuterVolumeSpecName: "kube-api-access-cxdtd") pod "d17d4141-7f2f-4598-9cf4-96870bca5903" (UID: "d17d4141-7f2f-4598-9cf4-96870bca5903"). InnerVolumeSpecName "kube-api-access-cxdtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.817951 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d17d4141-7f2f-4598-9cf4-96870bca5903" (UID: "d17d4141-7f2f-4598-9cf4-96870bca5903"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.837377 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-inventory" (OuterVolumeSpecName: "inventory") pod "d17d4141-7f2f-4598-9cf4-96870bca5903" (UID: "d17d4141-7f2f-4598-9cf4-96870bca5903"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.894148 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxdtd\" (UniqueName: \"kubernetes.io/projected/d17d4141-7f2f-4598-9cf4-96870bca5903-kube-api-access-cxdtd\") on node \"crc\" DevicePath \"\"" Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.894609 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:46:46 crc kubenswrapper[4876]: I1205 08:46:46.894676 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d17d4141-7f2f-4598-9cf4-96870bca5903-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.191196 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" event={"ID":"d17d4141-7f2f-4598-9cf4-96870bca5903","Type":"ContainerDied","Data":"b1b9a7ebb2267d033db3994bdf49644e8c9f93a54890c0001be262f24c7e87f2"} Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.191491 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1b9a7ebb2267d033db3994bdf49644e8c9f93a54890c0001be262f24c7e87f2" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.191464 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.274263 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln"] Dec 05 08:46:47 crc kubenswrapper[4876]: E1205 08:46:47.274675 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17d4141-7f2f-4598-9cf4-96870bca5903" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.274695 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17d4141-7f2f-4598-9cf4-96870bca5903" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 05 08:46:47 crc kubenswrapper[4876]: E1205 08:46:47.274721 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c" containerName="collect-profiles" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.274730 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c" containerName="collect-profiles" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.274965 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17d4141-7f2f-4598-9cf4-96870bca5903" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.274991 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c" containerName="collect-profiles" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.275653 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.277964 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.278139 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.278213 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.281245 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.291948 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln"] Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.403329 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdcrb\" (UniqueName: \"kubernetes.io/projected/9d51e73b-805b-4690-aed5-514789933cce-kube-api-access-rdcrb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.403591 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.403672 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.505245 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.505896 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.506250 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdcrb\" (UniqueName: \"kubernetes.io/projected/9d51e73b-805b-4690-aed5-514789933cce-kube-api-access-rdcrb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.509974 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.510140 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.528102 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdcrb\" (UniqueName: \"kubernetes.io/projected/9d51e73b-805b-4690-aed5-514789933cce-kube-api-access-rdcrb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-pfqln\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:47 crc kubenswrapper[4876]: I1205 08:46:47.593136 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:46:48 crc kubenswrapper[4876]: I1205 08:46:48.084667 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln"] Dec 05 08:46:48 crc kubenswrapper[4876]: W1205 08:46:48.089709 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d51e73b_805b_4690_aed5_514789933cce.slice/crio-e7fc218bc33aa9ca71b6e92e248dd1667e4527268a3f8a5395d007badb06b467 WatchSource:0}: Error finding container e7fc218bc33aa9ca71b6e92e248dd1667e4527268a3f8a5395d007badb06b467: Status 404 returned error can't find the container with id e7fc218bc33aa9ca71b6e92e248dd1667e4527268a3f8a5395d007badb06b467 Dec 05 08:46:48 crc kubenswrapper[4876]: I1205 08:46:48.208171 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" event={"ID":"9d51e73b-805b-4690-aed5-514789933cce","Type":"ContainerStarted","Data":"e7fc218bc33aa9ca71b6e92e248dd1667e4527268a3f8a5395d007badb06b467"} Dec 05 08:46:49 crc kubenswrapper[4876]: I1205 08:46:49.223050 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" event={"ID":"9d51e73b-805b-4690-aed5-514789933cce","Type":"ContainerStarted","Data":"9695cc05f6049e3f853c60a028a7b9a349a909ba2a494e7760737578a84a85fd"} Dec 05 08:46:49 crc kubenswrapper[4876]: I1205 08:46:49.249152 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" podStartSLOduration=1.836610185 podStartE2EDuration="2.24913424s" podCreationTimestamp="2025-12-05 08:46:47 +0000 UTC" firstStartedPulling="2025-12-05 08:46:48.094156546 +0000 UTC m=+1872.582821178" lastFinishedPulling="2025-12-05 08:46:48.506680611 +0000 UTC m=+1872.995345233" observedRunningTime="2025-12-05 08:46:49.247795523 +0000 UTC m=+1873.736460185" watchObservedRunningTime="2025-12-05 08:46:49.24913424 +0000 UTC m=+1873.737798862" Dec 05 08:47:17 crc kubenswrapper[4876]: I1205 08:47:17.055583 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-z22v8"] Dec 05 08:47:17 crc kubenswrapper[4876]: I1205 08:47:17.066386 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-87f7-account-create-update-fsgxj"] Dec 05 08:47:17 crc kubenswrapper[4876]: I1205 08:47:17.076373 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-87f7-account-create-update-fsgxj"] Dec 05 08:47:17 crc kubenswrapper[4876]: I1205 08:47:17.087683 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-z22v8"] Dec 05 08:47:17 crc kubenswrapper[4876]: I1205 08:47:17.834531 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b58b1a1-c690-4a1c-9280-09c5baa7a5a3" path="/var/lib/kubelet/pods/9b58b1a1-c690-4a1c-9280-09c5baa7a5a3/volumes" Dec 05 08:47:17 crc kubenswrapper[4876]: I1205 08:47:17.835181 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2bce654-ab4b-4687-8216-21761f83a696" path="/var/lib/kubelet/pods/d2bce654-ab4b-4687-8216-21761f83a696/volumes" Dec 05 08:47:18 crc kubenswrapper[4876]: I1205 08:47:18.034113 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hb5wk"] Dec 05 08:47:18 crc kubenswrapper[4876]: I1205 08:47:18.044631 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-9cljp"] Dec 05 08:47:18 crc kubenswrapper[4876]: I1205 08:47:18.056161 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-d661-account-create-update-cb7c4"] Dec 05 08:47:18 crc kubenswrapper[4876]: I1205 08:47:18.064702 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-b05e-account-create-update-5t4rh"] Dec 05 08:47:18 crc kubenswrapper[4876]: I1205 08:47:18.072386 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hb5wk"] Dec 05 08:47:18 crc kubenswrapper[4876]: I1205 08:47:18.079166 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-b05e-account-create-update-5t4rh"] Dec 05 08:47:18 crc kubenswrapper[4876]: I1205 08:47:18.107192 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-9cljp"] Dec 05 08:47:18 crc kubenswrapper[4876]: I1205 08:47:18.116290 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-d661-account-create-update-cb7c4"] Dec 05 08:47:19 crc kubenswrapper[4876]: I1205 08:47:19.840239 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e638147-a228-4f35-b3f1-8ead1cc5db48" path="/var/lib/kubelet/pods/1e638147-a228-4f35-b3f1-8ead1cc5db48/volumes" Dec 05 08:47:19 crc kubenswrapper[4876]: I1205 08:47:19.841380 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="693e7fd3-fc4e-44fe-827c-2063e9d2f496" path="/var/lib/kubelet/pods/693e7fd3-fc4e-44fe-827c-2063e9d2f496/volumes" Dec 05 08:47:19 crc kubenswrapper[4876]: I1205 08:47:19.842100 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed3eafa5-4131-4235-aba0-de93198af37d" path="/var/lib/kubelet/pods/ed3eafa5-4131-4235-aba0-de93198af37d/volumes" Dec 05 08:47:19 crc kubenswrapper[4876]: I1205 08:47:19.842709 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f70aa4ff-ac52-4cef-808e-6133ef7d395f" path="/var/lib/kubelet/pods/f70aa4ff-ac52-4cef-808e-6133ef7d395f/volumes" Dec 05 08:47:41 crc kubenswrapper[4876]: I1205 08:47:41.081827 4876 scope.go:117] "RemoveContainer" containerID="b405581d05fc048054062d7341eb0a64d71fe644de6c1d05cd459c513614b0ac" Dec 05 08:47:41 crc kubenswrapper[4876]: I1205 08:47:41.104111 4876 scope.go:117] "RemoveContainer" containerID="117e5225f0ecf9e60d6a447160902698a1ba249404e7f25f760ef4f1bb4fe123" Dec 05 08:47:41 crc kubenswrapper[4876]: I1205 08:47:41.154412 4876 scope.go:117] "RemoveContainer" containerID="5e0777b08a3d7a1ed3ab11d5a47a1bc7dbe631480df9c22897e474b120904b6f" Dec 05 08:47:41 crc kubenswrapper[4876]: I1205 08:47:41.252325 4876 scope.go:117] "RemoveContainer" containerID="16ca771999698bdf2bf33649c6cb9a6aa528f806ec5c1766673c609bc8d43e45" Dec 05 08:47:41 crc kubenswrapper[4876]: I1205 08:47:41.283649 4876 scope.go:117] "RemoveContainer" containerID="d2bffe9292c162769ac09a3e456a25f0559b47668f715d740b0bcedbd0ca85e9" Dec 05 08:47:41 crc kubenswrapper[4876]: I1205 08:47:41.352314 4876 scope.go:117] "RemoveContainer" containerID="a591cb44f594d6588a92f9089754f675bf3be8a1ab2a2a8a17c47ffa58952fbd" Dec 05 08:47:52 crc kubenswrapper[4876]: I1205 08:47:52.045561 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-54h52"] Dec 05 08:47:52 crc kubenswrapper[4876]: I1205 08:47:52.056828 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-54h52"] Dec 05 08:47:53 crc kubenswrapper[4876]: I1205 08:47:53.836240 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99d7f1cf-79c4-42c4-bd75-a6588e8daaec" path="/var/lib/kubelet/pods/99d7f1cf-79c4-42c4-bd75-a6588e8daaec/volumes" Dec 05 08:48:01 crc kubenswrapper[4876]: I1205 08:48:01.871174 4876 generic.go:334] "Generic (PLEG): container finished" podID="9d51e73b-805b-4690-aed5-514789933cce" containerID="9695cc05f6049e3f853c60a028a7b9a349a909ba2a494e7760737578a84a85fd" exitCode=0 Dec 05 08:48:01 crc kubenswrapper[4876]: I1205 08:48:01.871271 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" event={"ID":"9d51e73b-805b-4690-aed5-514789933cce","Type":"ContainerDied","Data":"9695cc05f6049e3f853c60a028a7b9a349a909ba2a494e7760737578a84a85fd"} Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.306410 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.462302 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-inventory\") pod \"9d51e73b-805b-4690-aed5-514789933cce\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.462682 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdcrb\" (UniqueName: \"kubernetes.io/projected/9d51e73b-805b-4690-aed5-514789933cce-kube-api-access-rdcrb\") pod \"9d51e73b-805b-4690-aed5-514789933cce\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.462746 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-ssh-key\") pod \"9d51e73b-805b-4690-aed5-514789933cce\" (UID: \"9d51e73b-805b-4690-aed5-514789933cce\") " Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.468243 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d51e73b-805b-4690-aed5-514789933cce-kube-api-access-rdcrb" (OuterVolumeSpecName: "kube-api-access-rdcrb") pod "9d51e73b-805b-4690-aed5-514789933cce" (UID: "9d51e73b-805b-4690-aed5-514789933cce"). InnerVolumeSpecName "kube-api-access-rdcrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.494096 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9d51e73b-805b-4690-aed5-514789933cce" (UID: "9d51e73b-805b-4690-aed5-514789933cce"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.518800 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-inventory" (OuterVolumeSpecName: "inventory") pod "9d51e73b-805b-4690-aed5-514789933cce" (UID: "9d51e73b-805b-4690-aed5-514789933cce"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.564840 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdcrb\" (UniqueName: \"kubernetes.io/projected/9d51e73b-805b-4690-aed5-514789933cce-kube-api-access-rdcrb\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.564880 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.564892 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d51e73b-805b-4690-aed5-514789933cce-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.891166 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" event={"ID":"9d51e73b-805b-4690-aed5-514789933cce","Type":"ContainerDied","Data":"e7fc218bc33aa9ca71b6e92e248dd1667e4527268a3f8a5395d007badb06b467"} Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.891209 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7fc218bc33aa9ca71b6e92e248dd1667e4527268a3f8a5395d007badb06b467" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.891209 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-pfqln" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.983224 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc"] Dec 05 08:48:03 crc kubenswrapper[4876]: E1205 08:48:03.983617 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d51e73b-805b-4690-aed5-514789933cce" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.983636 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d51e73b-805b-4690-aed5-514789933cce" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.983869 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d51e73b-805b-4690-aed5-514789933cce" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.984837 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.988657 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.989067 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.989280 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:48:03 crc kubenswrapper[4876]: I1205 08:48:03.990520 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.027407 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc"] Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.181861 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnfz4\" (UniqueName: \"kubernetes.io/projected/91a239cd-2756-4d03-8175-32dd507c0b2a-kube-api-access-tnfz4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.182308 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.182372 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.283147 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.283196 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.283312 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnfz4\" (UniqueName: \"kubernetes.io/projected/91a239cd-2756-4d03-8175-32dd507c0b2a-kube-api-access-tnfz4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.287034 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.291277 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.309320 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnfz4\" (UniqueName: \"kubernetes.io/projected/91a239cd-2756-4d03-8175-32dd507c0b2a-kube-api-access-tnfz4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-77bdc\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.317362 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.876196 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc"] Dec 05 08:48:04 crc kubenswrapper[4876]: I1205 08:48:04.903276 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" event={"ID":"91a239cd-2756-4d03-8175-32dd507c0b2a","Type":"ContainerStarted","Data":"508b4720cd94ac10ba0a26e1c49b6baa8dbf4fc728941658ff0f5c9edbcad9b4"} Dec 05 08:48:05 crc kubenswrapper[4876]: I1205 08:48:05.912523 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" event={"ID":"91a239cd-2756-4d03-8175-32dd507c0b2a","Type":"ContainerStarted","Data":"bbbe0424ecae56d57d8cd8ddc65a5823a551ba3f3092a4853c3715a05fbcbe4c"} Dec 05 08:48:05 crc kubenswrapper[4876]: I1205 08:48:05.931334 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" podStartSLOduration=2.432774711 podStartE2EDuration="2.931316354s" podCreationTimestamp="2025-12-05 08:48:03 +0000 UTC" firstStartedPulling="2025-12-05 08:48:04.884043151 +0000 UTC m=+1949.372707773" lastFinishedPulling="2025-12-05 08:48:05.382584784 +0000 UTC m=+1949.871249416" observedRunningTime="2025-12-05 08:48:05.930915422 +0000 UTC m=+1950.419580044" watchObservedRunningTime="2025-12-05 08:48:05.931316354 +0000 UTC m=+1950.419980976" Dec 05 08:48:08 crc kubenswrapper[4876]: I1205 08:48:08.213497 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:48:08 crc kubenswrapper[4876]: I1205 08:48:08.214089 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:48:10 crc kubenswrapper[4876]: I1205 08:48:10.964837 4876 generic.go:334] "Generic (PLEG): container finished" podID="91a239cd-2756-4d03-8175-32dd507c0b2a" containerID="bbbe0424ecae56d57d8cd8ddc65a5823a551ba3f3092a4853c3715a05fbcbe4c" exitCode=0 Dec 05 08:48:10 crc kubenswrapper[4876]: I1205 08:48:10.964947 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" event={"ID":"91a239cd-2756-4d03-8175-32dd507c0b2a","Type":"ContainerDied","Data":"bbbe0424ecae56d57d8cd8ddc65a5823a551ba3f3092a4853c3715a05fbcbe4c"} Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.390932 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.634877 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-inventory\") pod \"91a239cd-2756-4d03-8175-32dd507c0b2a\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.635102 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnfz4\" (UniqueName: \"kubernetes.io/projected/91a239cd-2756-4d03-8175-32dd507c0b2a-kube-api-access-tnfz4\") pod \"91a239cd-2756-4d03-8175-32dd507c0b2a\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.635169 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-ssh-key\") pod \"91a239cd-2756-4d03-8175-32dd507c0b2a\" (UID: \"91a239cd-2756-4d03-8175-32dd507c0b2a\") " Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.641732 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91a239cd-2756-4d03-8175-32dd507c0b2a-kube-api-access-tnfz4" (OuterVolumeSpecName: "kube-api-access-tnfz4") pod "91a239cd-2756-4d03-8175-32dd507c0b2a" (UID: "91a239cd-2756-4d03-8175-32dd507c0b2a"). InnerVolumeSpecName "kube-api-access-tnfz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.661127 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "91a239cd-2756-4d03-8175-32dd507c0b2a" (UID: "91a239cd-2756-4d03-8175-32dd507c0b2a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.667136 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-inventory" (OuterVolumeSpecName: "inventory") pod "91a239cd-2756-4d03-8175-32dd507c0b2a" (UID: "91a239cd-2756-4d03-8175-32dd507c0b2a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.737626 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.737653 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnfz4\" (UniqueName: \"kubernetes.io/projected/91a239cd-2756-4d03-8175-32dd507c0b2a-kube-api-access-tnfz4\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.737666 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91a239cd-2756-4d03-8175-32dd507c0b2a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.985341 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" event={"ID":"91a239cd-2756-4d03-8175-32dd507c0b2a","Type":"ContainerDied","Data":"508b4720cd94ac10ba0a26e1c49b6baa8dbf4fc728941658ff0f5c9edbcad9b4"} Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.985815 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="508b4720cd94ac10ba0a26e1c49b6baa8dbf4fc728941658ff0f5c9edbcad9b4" Dec 05 08:48:12 crc kubenswrapper[4876]: I1205 08:48:12.985398 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-77bdc" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.073717 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72"] Dec 05 08:48:13 crc kubenswrapper[4876]: E1205 08:48:13.074336 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a239cd-2756-4d03-8175-32dd507c0b2a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.074363 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a239cd-2756-4d03-8175-32dd507c0b2a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.074607 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="91a239cd-2756-4d03-8175-32dd507c0b2a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.075686 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.079301 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.079416 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.079298 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.080173 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.092067 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72"] Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.150604 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.150675 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.150803 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz5cp\" (UniqueName: \"kubernetes.io/projected/f596bae7-c567-469e-b7be-f95265c0f016-kube-api-access-cz5cp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.253215 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.253267 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz5cp\" (UniqueName: \"kubernetes.io/projected/f596bae7-c567-469e-b7be-f95265c0f016-kube-api-access-cz5cp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.253426 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.257705 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.258345 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.270055 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz5cp\" (UniqueName: \"kubernetes.io/projected/f596bae7-c567-469e-b7be-f95265c0f016-kube-api-access-cz5cp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-slf72\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.395311 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:13 crc kubenswrapper[4876]: W1205 08:48:13.949547 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf596bae7_c567_469e_b7be_f95265c0f016.slice/crio-78df2a24d5ef809c4326984f81ffdd8cf8c3d40c0f798958af7456822d637a35 WatchSource:0}: Error finding container 78df2a24d5ef809c4326984f81ffdd8cf8c3d40c0f798958af7456822d637a35: Status 404 returned error can't find the container with id 78df2a24d5ef809c4326984f81ffdd8cf8c3d40c0f798958af7456822d637a35 Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.949658 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72"] Dec 05 08:48:13 crc kubenswrapper[4876]: I1205 08:48:13.996379 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" event={"ID":"f596bae7-c567-469e-b7be-f95265c0f016","Type":"ContainerStarted","Data":"78df2a24d5ef809c4326984f81ffdd8cf8c3d40c0f798958af7456822d637a35"} Dec 05 08:48:15 crc kubenswrapper[4876]: I1205 08:48:15.006417 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" event={"ID":"f596bae7-c567-469e-b7be-f95265c0f016","Type":"ContainerStarted","Data":"1fc1b2e6c160a9ec9464abef1420eecae5c0c0a3590631d1ee23a440404902e7"} Dec 05 08:48:16 crc kubenswrapper[4876]: I1205 08:48:16.040891 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" podStartSLOduration=2.638399716 podStartE2EDuration="3.04086569s" podCreationTimestamp="2025-12-05 08:48:13 +0000 UTC" firstStartedPulling="2025-12-05 08:48:13.951923214 +0000 UTC m=+1958.440587836" lastFinishedPulling="2025-12-05 08:48:14.354389188 +0000 UTC m=+1958.843053810" observedRunningTime="2025-12-05 08:48:15.034836199 +0000 UTC m=+1959.523500831" watchObservedRunningTime="2025-12-05 08:48:16.04086569 +0000 UTC m=+1960.529530322" Dec 05 08:48:16 crc kubenswrapper[4876]: I1205 08:48:16.044921 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vhxdf"] Dec 05 08:48:16 crc kubenswrapper[4876]: I1205 08:48:16.053862 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vhxdf"] Dec 05 08:48:17 crc kubenswrapper[4876]: I1205 08:48:17.031333 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gpm57"] Dec 05 08:48:17 crc kubenswrapper[4876]: I1205 08:48:17.041747 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gpm57"] Dec 05 08:48:17 crc kubenswrapper[4876]: I1205 08:48:17.836632 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="198b165f-0f97-4551-a939-73d07f958d01" path="/var/lib/kubelet/pods/198b165f-0f97-4551-a939-73d07f958d01/volumes" Dec 05 08:48:17 crc kubenswrapper[4876]: I1205 08:48:17.837771 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6629f58-7171-4cd4-a018-6b9c40a8bfee" path="/var/lib/kubelet/pods/b6629f58-7171-4cd4-a018-6b9c40a8bfee/volumes" Dec 05 08:48:38 crc kubenswrapper[4876]: I1205 08:48:38.214271 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:48:38 crc kubenswrapper[4876]: I1205 08:48:38.216184 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:48:40 crc kubenswrapper[4876]: I1205 08:48:40.041946 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-n696k"] Dec 05 08:48:40 crc kubenswrapper[4876]: I1205 08:48:40.055637 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-n696k"] Dec 05 08:48:41 crc kubenswrapper[4876]: I1205 08:48:41.533307 4876 scope.go:117] "RemoveContainer" containerID="fcc9274310b57ba2fa1cba5b1f467a6e7fe5d2ca5ed7ea581df54d9d31ce767c" Dec 05 08:48:41 crc kubenswrapper[4876]: I1205 08:48:41.605326 4876 scope.go:117] "RemoveContainer" containerID="f87765f45920f1ea4498d1234b240f24872ab33ead370a56069f6fe647c0531e" Dec 05 08:48:41 crc kubenswrapper[4876]: I1205 08:48:41.666021 4876 scope.go:117] "RemoveContainer" containerID="61ca35015ec9ca00cca71b6d6bd2a6cd1722b775213b9192131b407f5798c455" Dec 05 08:48:41 crc kubenswrapper[4876]: I1205 08:48:41.836714 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36f2a90b-0c85-478b-83e5-17ff10e01c94" path="/var/lib/kubelet/pods/36f2a90b-0c85-478b-83e5-17ff10e01c94/volumes" Dec 05 08:48:52 crc kubenswrapper[4876]: I1205 08:48:52.390716 4876 generic.go:334] "Generic (PLEG): container finished" podID="f596bae7-c567-469e-b7be-f95265c0f016" containerID="1fc1b2e6c160a9ec9464abef1420eecae5c0c0a3590631d1ee23a440404902e7" exitCode=0 Dec 05 08:48:52 crc kubenswrapper[4876]: I1205 08:48:52.390820 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" event={"ID":"f596bae7-c567-469e-b7be-f95265c0f016","Type":"ContainerDied","Data":"1fc1b2e6c160a9ec9464abef1420eecae5c0c0a3590631d1ee23a440404902e7"} Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.806323 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.865069 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz5cp\" (UniqueName: \"kubernetes.io/projected/f596bae7-c567-469e-b7be-f95265c0f016-kube-api-access-cz5cp\") pod \"f596bae7-c567-469e-b7be-f95265c0f016\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.865428 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-inventory\") pod \"f596bae7-c567-469e-b7be-f95265c0f016\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.865538 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-ssh-key\") pod \"f596bae7-c567-469e-b7be-f95265c0f016\" (UID: \"f596bae7-c567-469e-b7be-f95265c0f016\") " Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.870375 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f596bae7-c567-469e-b7be-f95265c0f016-kube-api-access-cz5cp" (OuterVolumeSpecName: "kube-api-access-cz5cp") pod "f596bae7-c567-469e-b7be-f95265c0f016" (UID: "f596bae7-c567-469e-b7be-f95265c0f016"). InnerVolumeSpecName "kube-api-access-cz5cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.894952 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-inventory" (OuterVolumeSpecName: "inventory") pod "f596bae7-c567-469e-b7be-f95265c0f016" (UID: "f596bae7-c567-469e-b7be-f95265c0f016"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.900568 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f596bae7-c567-469e-b7be-f95265c0f016" (UID: "f596bae7-c567-469e-b7be-f95265c0f016"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.967840 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz5cp\" (UniqueName: \"kubernetes.io/projected/f596bae7-c567-469e-b7be-f95265c0f016-kube-api-access-cz5cp\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.967887 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:53 crc kubenswrapper[4876]: I1205 08:48:53.967906 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f596bae7-c567-469e-b7be-f95265c0f016-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.418660 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" event={"ID":"f596bae7-c567-469e-b7be-f95265c0f016","Type":"ContainerDied","Data":"78df2a24d5ef809c4326984f81ffdd8cf8c3d40c0f798958af7456822d637a35"} Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.419058 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78df2a24d5ef809c4326984f81ffdd8cf8c3d40c0f798958af7456822d637a35" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.418773 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-slf72" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.502392 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z"] Dec 05 08:48:54 crc kubenswrapper[4876]: E1205 08:48:54.503086 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f596bae7-c567-469e-b7be-f95265c0f016" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.503114 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f596bae7-c567-469e-b7be-f95265c0f016" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.503378 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f596bae7-c567-469e-b7be-f95265c0f016" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.504335 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.508597 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.509103 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.509281 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.509497 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.519247 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z"] Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.680044 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.680307 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.680603 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmzw5\" (UniqueName: \"kubernetes.io/projected/8a524784-c668-4388-81c5-719f97dbb947-kube-api-access-dmzw5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.781723 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmzw5\" (UniqueName: \"kubernetes.io/projected/8a524784-c668-4388-81c5-719f97dbb947-kube-api-access-dmzw5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.781824 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.781881 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.786436 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.787813 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.805296 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmzw5\" (UniqueName: \"kubernetes.io/projected/8a524784-c668-4388-81c5-719f97dbb947-kube-api-access-dmzw5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:54 crc kubenswrapper[4876]: I1205 08:48:54.840161 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:48:55 crc kubenswrapper[4876]: I1205 08:48:55.228466 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z"] Dec 05 08:48:55 crc kubenswrapper[4876]: I1205 08:48:55.427438 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" event={"ID":"8a524784-c668-4388-81c5-719f97dbb947","Type":"ContainerStarted","Data":"29cfe61295178aff155aa1dbe81dfc3bc5644a0b66fd37079b83dc7db0f9db1e"} Dec 05 08:48:56 crc kubenswrapper[4876]: I1205 08:48:56.434824 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" event={"ID":"8a524784-c668-4388-81c5-719f97dbb947","Type":"ContainerStarted","Data":"77c2d3db7aece69277c5800ecaee9785f8ee75a5e2c496976e430c9a5fff058b"} Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.186677 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" podStartSLOduration=6.734660304 podStartE2EDuration="7.186655505s" podCreationTimestamp="2025-12-05 08:48:54 +0000 UTC" firstStartedPulling="2025-12-05 08:48:55.231414892 +0000 UTC m=+1999.720079504" lastFinishedPulling="2025-12-05 08:48:55.683410083 +0000 UTC m=+2000.172074705" observedRunningTime="2025-12-05 08:48:56.450427579 +0000 UTC m=+2000.939092201" watchObservedRunningTime="2025-12-05 08:49:01.186655505 +0000 UTC m=+2005.675320127" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.193299 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fl945"] Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.196052 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.200180 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-utilities\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.200253 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-475kh\" (UniqueName: \"kubernetes.io/projected/bc1e414e-6375-4382-b1e0-17c4cff84425-kube-api-access-475kh\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.200318 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-catalog-content\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.206937 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fl945"] Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.301437 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-catalog-content\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.301571 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-utilities\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.301628 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-475kh\" (UniqueName: \"kubernetes.io/projected/bc1e414e-6375-4382-b1e0-17c4cff84425-kube-api-access-475kh\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.302574 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-catalog-content\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.302645 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-utilities\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.322574 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-475kh\" (UniqueName: \"kubernetes.io/projected/bc1e414e-6375-4382-b1e0-17c4cff84425-kube-api-access-475kh\") pod \"community-operators-fl945\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:01 crc kubenswrapper[4876]: I1205 08:49:01.518880 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:02 crc kubenswrapper[4876]: I1205 08:49:02.105854 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fl945"] Dec 05 08:49:02 crc kubenswrapper[4876]: I1205 08:49:02.486497 4876 generic.go:334] "Generic (PLEG): container finished" podID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerID="698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e" exitCode=0 Dec 05 08:49:02 crc kubenswrapper[4876]: I1205 08:49:02.486601 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl945" event={"ID":"bc1e414e-6375-4382-b1e0-17c4cff84425","Type":"ContainerDied","Data":"698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e"} Dec 05 08:49:02 crc kubenswrapper[4876]: I1205 08:49:02.486773 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl945" event={"ID":"bc1e414e-6375-4382-b1e0-17c4cff84425","Type":"ContainerStarted","Data":"f768d7dbbaa16eb916a56ed09555d405e2872e009ed231ead81c27983803f886"} Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:03.999955 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t8kzk"] Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.002239 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.017582 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t8kzk"] Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.173250 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-utilities\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.173364 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-catalog-content\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.173764 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rvmx\" (UniqueName: \"kubernetes.io/projected/15ef4d9c-2dcd-40b1-83a4-9983b6054965-kube-api-access-6rvmx\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.276315 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-utilities\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.276398 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-catalog-content\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.276590 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rvmx\" (UniqueName: \"kubernetes.io/projected/15ef4d9c-2dcd-40b1-83a4-9983b6054965-kube-api-access-6rvmx\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.276827 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-utilities\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.276882 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-catalog-content\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.298829 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rvmx\" (UniqueName: \"kubernetes.io/projected/15ef4d9c-2dcd-40b1-83a4-9983b6054965-kube-api-access-6rvmx\") pod \"redhat-operators-t8kzk\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.320757 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.510283 4876 generic.go:334] "Generic (PLEG): container finished" podID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerID="616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f" exitCode=0 Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.510491 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl945" event={"ID":"bc1e414e-6375-4382-b1e0-17c4cff84425","Type":"ContainerDied","Data":"616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f"} Dec 05 08:49:04 crc kubenswrapper[4876]: I1205 08:49:04.793802 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t8kzk"] Dec 05 08:49:04 crc kubenswrapper[4876]: W1205 08:49:04.797062 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15ef4d9c_2dcd_40b1_83a4_9983b6054965.slice/crio-4df5fd6e001adb4e97b91943021ce8b939e1ebb6b34556b0f5190f280e540cbd WatchSource:0}: Error finding container 4df5fd6e001adb4e97b91943021ce8b939e1ebb6b34556b0f5190f280e540cbd: Status 404 returned error can't find the container with id 4df5fd6e001adb4e97b91943021ce8b939e1ebb6b34556b0f5190f280e540cbd Dec 05 08:49:05 crc kubenswrapper[4876]: I1205 08:49:05.520505 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl945" event={"ID":"bc1e414e-6375-4382-b1e0-17c4cff84425","Type":"ContainerStarted","Data":"a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594"} Dec 05 08:49:05 crc kubenswrapper[4876]: I1205 08:49:05.522358 4876 generic.go:334] "Generic (PLEG): container finished" podID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerID="4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0" exitCode=0 Dec 05 08:49:05 crc kubenswrapper[4876]: I1205 08:49:05.522393 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8kzk" event={"ID":"15ef4d9c-2dcd-40b1-83a4-9983b6054965","Type":"ContainerDied","Data":"4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0"} Dec 05 08:49:05 crc kubenswrapper[4876]: I1205 08:49:05.522424 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8kzk" event={"ID":"15ef4d9c-2dcd-40b1-83a4-9983b6054965","Type":"ContainerStarted","Data":"4df5fd6e001adb4e97b91943021ce8b939e1ebb6b34556b0f5190f280e540cbd"} Dec 05 08:49:05 crc kubenswrapper[4876]: I1205 08:49:05.544205 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fl945" podStartSLOduration=2.048262432 podStartE2EDuration="4.544190799s" podCreationTimestamp="2025-12-05 08:49:01 +0000 UTC" firstStartedPulling="2025-12-05 08:49:02.488188581 +0000 UTC m=+2006.976853203" lastFinishedPulling="2025-12-05 08:49:04.984116948 +0000 UTC m=+2009.472781570" observedRunningTime="2025-12-05 08:49:05.542056154 +0000 UTC m=+2010.030720766" watchObservedRunningTime="2025-12-05 08:49:05.544190799 +0000 UTC m=+2010.032855421" Dec 05 08:49:06 crc kubenswrapper[4876]: I1205 08:49:06.534618 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8kzk" event={"ID":"15ef4d9c-2dcd-40b1-83a4-9983b6054965","Type":"ContainerStarted","Data":"0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575"} Dec 05 08:49:08 crc kubenswrapper[4876]: I1205 08:49:08.213418 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:49:08 crc kubenswrapper[4876]: I1205 08:49:08.213799 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:49:08 crc kubenswrapper[4876]: I1205 08:49:08.213874 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:49:08 crc kubenswrapper[4876]: I1205 08:49:08.215316 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ca3218eccf9f7164d33de275ebcdaf5bcf4922b876f83c58de4aed55797c0a8"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:49:08 crc kubenswrapper[4876]: I1205 08:49:08.216008 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://5ca3218eccf9f7164d33de275ebcdaf5bcf4922b876f83c58de4aed55797c0a8" gracePeriod=600 Dec 05 08:49:08 crc kubenswrapper[4876]: I1205 08:49:08.552573 4876 generic.go:334] "Generic (PLEG): container finished" podID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerID="0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575" exitCode=0 Dec 05 08:49:08 crc kubenswrapper[4876]: I1205 08:49:08.552648 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8kzk" event={"ID":"15ef4d9c-2dcd-40b1-83a4-9983b6054965","Type":"ContainerDied","Data":"0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575"} Dec 05 08:49:10 crc kubenswrapper[4876]: I1205 08:49:10.578838 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="5ca3218eccf9f7164d33de275ebcdaf5bcf4922b876f83c58de4aed55797c0a8" exitCode=0 Dec 05 08:49:10 crc kubenswrapper[4876]: I1205 08:49:10.578955 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"5ca3218eccf9f7164d33de275ebcdaf5bcf4922b876f83c58de4aed55797c0a8"} Dec 05 08:49:10 crc kubenswrapper[4876]: I1205 08:49:10.579310 4876 scope.go:117] "RemoveContainer" containerID="4c58217de9b6d388046f4720b49da84c7255fe6a91b3bb4ed06f8ae3173f26d8" Dec 05 08:49:11 crc kubenswrapper[4876]: I1205 08:49:11.519137 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:11 crc kubenswrapper[4876]: I1205 08:49:11.519883 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:11 crc kubenswrapper[4876]: I1205 08:49:11.604161 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8kzk" event={"ID":"15ef4d9c-2dcd-40b1-83a4-9983b6054965","Type":"ContainerStarted","Data":"e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d"} Dec 05 08:49:11 crc kubenswrapper[4876]: I1205 08:49:11.608921 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412"} Dec 05 08:49:11 crc kubenswrapper[4876]: I1205 08:49:11.630317 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:11 crc kubenswrapper[4876]: I1205 08:49:11.670859 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t8kzk" podStartSLOduration=3.665779198 podStartE2EDuration="8.670841883s" podCreationTimestamp="2025-12-05 08:49:03 +0000 UTC" firstStartedPulling="2025-12-05 08:49:05.523755111 +0000 UTC m=+2010.012419743" lastFinishedPulling="2025-12-05 08:49:10.528817796 +0000 UTC m=+2015.017482428" observedRunningTime="2025-12-05 08:49:11.639145874 +0000 UTC m=+2016.127810496" watchObservedRunningTime="2025-12-05 08:49:11.670841883 +0000 UTC m=+2016.159506505" Dec 05 08:49:11 crc kubenswrapper[4876]: I1205 08:49:11.677485 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:12 crc kubenswrapper[4876]: I1205 08:49:12.984124 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fl945"] Dec 05 08:49:13 crc kubenswrapper[4876]: I1205 08:49:13.622079 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fl945" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerName="registry-server" containerID="cri-o://a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594" gracePeriod=2 Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.089595 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.284310 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-utilities\") pod \"bc1e414e-6375-4382-b1e0-17c4cff84425\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.284472 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-475kh\" (UniqueName: \"kubernetes.io/projected/bc1e414e-6375-4382-b1e0-17c4cff84425-kube-api-access-475kh\") pod \"bc1e414e-6375-4382-b1e0-17c4cff84425\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.284537 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-catalog-content\") pod \"bc1e414e-6375-4382-b1e0-17c4cff84425\" (UID: \"bc1e414e-6375-4382-b1e0-17c4cff84425\") " Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.285049 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-utilities" (OuterVolumeSpecName: "utilities") pod "bc1e414e-6375-4382-b1e0-17c4cff84425" (UID: "bc1e414e-6375-4382-b1e0-17c4cff84425"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.292121 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc1e414e-6375-4382-b1e0-17c4cff84425-kube-api-access-475kh" (OuterVolumeSpecName: "kube-api-access-475kh") pod "bc1e414e-6375-4382-b1e0-17c4cff84425" (UID: "bc1e414e-6375-4382-b1e0-17c4cff84425"). InnerVolumeSpecName "kube-api-access-475kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.321390 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.321444 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.341409 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc1e414e-6375-4382-b1e0-17c4cff84425" (UID: "bc1e414e-6375-4382-b1e0-17c4cff84425"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.387021 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.387061 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc1e414e-6375-4382-b1e0-17c4cff84425-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.387076 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-475kh\" (UniqueName: \"kubernetes.io/projected/bc1e414e-6375-4382-b1e0-17c4cff84425-kube-api-access-475kh\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.634271 4876 generic.go:334] "Generic (PLEG): container finished" podID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerID="a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594" exitCode=0 Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.634339 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fl945" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.634354 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl945" event={"ID":"bc1e414e-6375-4382-b1e0-17c4cff84425","Type":"ContainerDied","Data":"a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594"} Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.635073 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fl945" event={"ID":"bc1e414e-6375-4382-b1e0-17c4cff84425","Type":"ContainerDied","Data":"f768d7dbbaa16eb916a56ed09555d405e2872e009ed231ead81c27983803f886"} Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.635124 4876 scope.go:117] "RemoveContainer" containerID="a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.659112 4876 scope.go:117] "RemoveContainer" containerID="616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.675423 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fl945"] Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.682249 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fl945"] Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.703186 4876 scope.go:117] "RemoveContainer" containerID="698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.732658 4876 scope.go:117] "RemoveContainer" containerID="a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594" Dec 05 08:49:14 crc kubenswrapper[4876]: E1205 08:49:14.733525 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594\": container with ID starting with a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594 not found: ID does not exist" containerID="a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.733588 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594"} err="failed to get container status \"a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594\": rpc error: code = NotFound desc = could not find container \"a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594\": container with ID starting with a0d38a84aa1ce7f2ae6488aa7b6ef0354bc26911926d2147c8ad93262c2fa594 not found: ID does not exist" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.733691 4876 scope.go:117] "RemoveContainer" containerID="616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f" Dec 05 08:49:14 crc kubenswrapper[4876]: E1205 08:49:14.734278 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f\": container with ID starting with 616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f not found: ID does not exist" containerID="616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.734329 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f"} err="failed to get container status \"616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f\": rpc error: code = NotFound desc = could not find container \"616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f\": container with ID starting with 616f1b182801b99cffe9e6b541680f431ca9ab501192e825f7587b9ce0c25c0f not found: ID does not exist" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.734365 4876 scope.go:117] "RemoveContainer" containerID="698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e" Dec 05 08:49:14 crc kubenswrapper[4876]: E1205 08:49:14.734820 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e\": container with ID starting with 698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e not found: ID does not exist" containerID="698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e" Dec 05 08:49:14 crc kubenswrapper[4876]: I1205 08:49:14.734878 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e"} err="failed to get container status \"698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e\": rpc error: code = NotFound desc = could not find container \"698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e\": container with ID starting with 698361b7b8c7a697b5c046827887744265a67d6e6c6df467e7715bf46922307e not found: ID does not exist" Dec 05 08:49:15 crc kubenswrapper[4876]: I1205 08:49:15.372185 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t8kzk" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="registry-server" probeResult="failure" output=< Dec 05 08:49:15 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 08:49:15 crc kubenswrapper[4876]: > Dec 05 08:49:15 crc kubenswrapper[4876]: I1205 08:49:15.849375 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" path="/var/lib/kubelet/pods/bc1e414e-6375-4382-b1e0-17c4cff84425/volumes" Dec 05 08:49:24 crc kubenswrapper[4876]: I1205 08:49:24.371393 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:24 crc kubenswrapper[4876]: I1205 08:49:24.424852 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:24 crc kubenswrapper[4876]: I1205 08:49:24.610291 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t8kzk"] Dec 05 08:49:25 crc kubenswrapper[4876]: I1205 08:49:25.753274 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t8kzk" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="registry-server" containerID="cri-o://e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d" gracePeriod=2 Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.193640 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.329998 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rvmx\" (UniqueName: \"kubernetes.io/projected/15ef4d9c-2dcd-40b1-83a4-9983b6054965-kube-api-access-6rvmx\") pod \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.330170 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-catalog-content\") pod \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.330240 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-utilities\") pod \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\" (UID: \"15ef4d9c-2dcd-40b1-83a4-9983b6054965\") " Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.331012 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-utilities" (OuterVolumeSpecName: "utilities") pod "15ef4d9c-2dcd-40b1-83a4-9983b6054965" (UID: "15ef4d9c-2dcd-40b1-83a4-9983b6054965"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.332038 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.337093 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15ef4d9c-2dcd-40b1-83a4-9983b6054965-kube-api-access-6rvmx" (OuterVolumeSpecName: "kube-api-access-6rvmx") pod "15ef4d9c-2dcd-40b1-83a4-9983b6054965" (UID: "15ef4d9c-2dcd-40b1-83a4-9983b6054965"). InnerVolumeSpecName "kube-api-access-6rvmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.434129 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rvmx\" (UniqueName: \"kubernetes.io/projected/15ef4d9c-2dcd-40b1-83a4-9983b6054965-kube-api-access-6rvmx\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.458802 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15ef4d9c-2dcd-40b1-83a4-9983b6054965" (UID: "15ef4d9c-2dcd-40b1-83a4-9983b6054965"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.535469 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15ef4d9c-2dcd-40b1-83a4-9983b6054965-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.764001 4876 generic.go:334] "Generic (PLEG): container finished" podID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerID="e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d" exitCode=0 Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.764064 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t8kzk" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.764087 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8kzk" event={"ID":"15ef4d9c-2dcd-40b1-83a4-9983b6054965","Type":"ContainerDied","Data":"e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d"} Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.765198 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t8kzk" event={"ID":"15ef4d9c-2dcd-40b1-83a4-9983b6054965","Type":"ContainerDied","Data":"4df5fd6e001adb4e97b91943021ce8b939e1ebb6b34556b0f5190f280e540cbd"} Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.765234 4876 scope.go:117] "RemoveContainer" containerID="e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.825651 4876 scope.go:117] "RemoveContainer" containerID="0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.843005 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t8kzk"] Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.852993 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t8kzk"] Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.861790 4876 scope.go:117] "RemoveContainer" containerID="4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.895937 4876 scope.go:117] "RemoveContainer" containerID="e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d" Dec 05 08:49:26 crc kubenswrapper[4876]: E1205 08:49:26.896294 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d\": container with ID starting with e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d not found: ID does not exist" containerID="e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.896338 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d"} err="failed to get container status \"e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d\": rpc error: code = NotFound desc = could not find container \"e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d\": container with ID starting with e50af983073cdeef47c5398dc7e8960f17583d75800639b0f84f8549bc5a3e1d not found: ID does not exist" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.896457 4876 scope.go:117] "RemoveContainer" containerID="0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575" Dec 05 08:49:26 crc kubenswrapper[4876]: E1205 08:49:26.896999 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575\": container with ID starting with 0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575 not found: ID does not exist" containerID="0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.897149 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575"} err="failed to get container status \"0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575\": rpc error: code = NotFound desc = could not find container \"0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575\": container with ID starting with 0855c328a5dd16555634bac255f02a8e9b73be8476d4d4d5f63c4d97fde80575 not found: ID does not exist" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.897167 4876 scope.go:117] "RemoveContainer" containerID="4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0" Dec 05 08:49:26 crc kubenswrapper[4876]: E1205 08:49:26.897414 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0\": container with ID starting with 4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0 not found: ID does not exist" containerID="4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0" Dec 05 08:49:26 crc kubenswrapper[4876]: I1205 08:49:26.897470 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0"} err="failed to get container status \"4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0\": rpc error: code = NotFound desc = could not find container \"4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0\": container with ID starting with 4c92ad233906f23ecec8efc6f60daa225917c16a4e2e62d8cef6c94a02e7b7e0 not found: ID does not exist" Dec 05 08:49:27 crc kubenswrapper[4876]: I1205 08:49:27.833884 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" path="/var/lib/kubelet/pods/15ef4d9c-2dcd-40b1-83a4-9983b6054965/volumes" Dec 05 08:49:41 crc kubenswrapper[4876]: I1205 08:49:41.771462 4876 scope.go:117] "RemoveContainer" containerID="3eabe4487bdee6ddd8756e2c4aec68ac32a724d478587a901ac484e08b44507e" Dec 05 08:49:44 crc kubenswrapper[4876]: I1205 08:49:44.938129 4876 generic.go:334] "Generic (PLEG): container finished" podID="8a524784-c668-4388-81c5-719f97dbb947" containerID="77c2d3db7aece69277c5800ecaee9785f8ee75a5e2c496976e430c9a5fff058b" exitCode=0 Dec 05 08:49:44 crc kubenswrapper[4876]: I1205 08:49:44.938252 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" event={"ID":"8a524784-c668-4388-81c5-719f97dbb947","Type":"ContainerDied","Data":"77c2d3db7aece69277c5800ecaee9785f8ee75a5e2c496976e430c9a5fff058b"} Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.346227 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.365036 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-ssh-key\") pod \"8a524784-c668-4388-81c5-719f97dbb947\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.365125 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-inventory\") pod \"8a524784-c668-4388-81c5-719f97dbb947\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.365329 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmzw5\" (UniqueName: \"kubernetes.io/projected/8a524784-c668-4388-81c5-719f97dbb947-kube-api-access-dmzw5\") pod \"8a524784-c668-4388-81c5-719f97dbb947\" (UID: \"8a524784-c668-4388-81c5-719f97dbb947\") " Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.384791 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a524784-c668-4388-81c5-719f97dbb947-kube-api-access-dmzw5" (OuterVolumeSpecName: "kube-api-access-dmzw5") pod "8a524784-c668-4388-81c5-719f97dbb947" (UID: "8a524784-c668-4388-81c5-719f97dbb947"). InnerVolumeSpecName "kube-api-access-dmzw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.406677 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a524784-c668-4388-81c5-719f97dbb947" (UID: "8a524784-c668-4388-81c5-719f97dbb947"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.409287 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-inventory" (OuterVolumeSpecName: "inventory") pod "8a524784-c668-4388-81c5-719f97dbb947" (UID: "8a524784-c668-4388-81c5-719f97dbb947"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.468066 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmzw5\" (UniqueName: \"kubernetes.io/projected/8a524784-c668-4388-81c5-719f97dbb947-kube-api-access-dmzw5\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.468093 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.468102 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a524784-c668-4388-81c5-719f97dbb947-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.958287 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" event={"ID":"8a524784-c668-4388-81c5-719f97dbb947","Type":"ContainerDied","Data":"29cfe61295178aff155aa1dbe81dfc3bc5644a0b66fd37079b83dc7db0f9db1e"} Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.958328 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29cfe61295178aff155aa1dbe81dfc3bc5644a0b66fd37079b83dc7db0f9db1e" Dec 05 08:49:46 crc kubenswrapper[4876]: I1205 08:49:46.958344 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.067576 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-stmmf"] Dec 05 08:49:47 crc kubenswrapper[4876]: E1205 08:49:47.067981 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a524784-c668-4388-81c5-719f97dbb947" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.067997 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a524784-c668-4388-81c5-719f97dbb947" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:49:47 crc kubenswrapper[4876]: E1205 08:49:47.068014 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerName="registry-server" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068021 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerName="registry-server" Dec 05 08:49:47 crc kubenswrapper[4876]: E1205 08:49:47.068035 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="extract-utilities" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068041 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="extract-utilities" Dec 05 08:49:47 crc kubenswrapper[4876]: E1205 08:49:47.068054 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="extract-content" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068060 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="extract-content" Dec 05 08:49:47 crc kubenswrapper[4876]: E1205 08:49:47.068068 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="registry-server" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068075 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="registry-server" Dec 05 08:49:47 crc kubenswrapper[4876]: E1205 08:49:47.068111 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerName="extract-content" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068119 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerName="extract-content" Dec 05 08:49:47 crc kubenswrapper[4876]: E1205 08:49:47.068131 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerName="extract-utilities" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068139 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerName="extract-utilities" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068380 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc1e414e-6375-4382-b1e0-17c4cff84425" containerName="registry-server" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068407 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="15ef4d9c-2dcd-40b1-83a4-9983b6054965" containerName="registry-server" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.068414 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a524784-c668-4388-81c5-719f97dbb947" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.069036 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.071186 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.071590 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.071746 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.071922 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.081046 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvdk6\" (UniqueName: \"kubernetes.io/projected/dc2909c6-f489-467c-a203-01adb27428f5-kube-api-access-pvdk6\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.081134 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.081175 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.093278 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-stmmf"] Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.183579 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvdk6\" (UniqueName: \"kubernetes.io/projected/dc2909c6-f489-467c-a203-01adb27428f5-kube-api-access-pvdk6\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.183729 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.183852 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.187688 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.187970 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.202052 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvdk6\" (UniqueName: \"kubernetes.io/projected/dc2909c6-f489-467c-a203-01adb27428f5-kube-api-access-pvdk6\") pod \"ssh-known-hosts-edpm-deployment-stmmf\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:47 crc kubenswrapper[4876]: I1205 08:49:47.392305 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:48 crc kubenswrapper[4876]: I1205 08:49:48.059452 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-stmmf"] Dec 05 08:49:48 crc kubenswrapper[4876]: I1205 08:49:48.988346 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" event={"ID":"dc2909c6-f489-467c-a203-01adb27428f5","Type":"ContainerStarted","Data":"0ff02679f36fc5087b174e3e94031ec1811cce5d535d64bfb619e528ce1075c5"} Dec 05 08:49:48 crc kubenswrapper[4876]: I1205 08:49:48.988429 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" event={"ID":"dc2909c6-f489-467c-a203-01adb27428f5","Type":"ContainerStarted","Data":"cddccff364d8a1ada5ca961edcde7e51065605b667703c08d12488bf02b83bd6"} Dec 05 08:49:49 crc kubenswrapper[4876]: I1205 08:49:49.095357 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" podStartSLOduration=1.6782108930000001 podStartE2EDuration="2.095337248s" podCreationTimestamp="2025-12-05 08:49:47 +0000 UTC" firstStartedPulling="2025-12-05 08:49:48.061578129 +0000 UTC m=+2052.550242751" lastFinishedPulling="2025-12-05 08:49:48.478704474 +0000 UTC m=+2052.967369106" observedRunningTime="2025-12-05 08:49:49.091349254 +0000 UTC m=+2053.580013876" watchObservedRunningTime="2025-12-05 08:49:49.095337248 +0000 UTC m=+2053.584001870" Dec 05 08:49:58 crc kubenswrapper[4876]: I1205 08:49:58.089706 4876 generic.go:334] "Generic (PLEG): container finished" podID="dc2909c6-f489-467c-a203-01adb27428f5" containerID="0ff02679f36fc5087b174e3e94031ec1811cce5d535d64bfb619e528ce1075c5" exitCode=0 Dec 05 08:49:58 crc kubenswrapper[4876]: I1205 08:49:58.089796 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" event={"ID":"dc2909c6-f489-467c-a203-01adb27428f5","Type":"ContainerDied","Data":"0ff02679f36fc5087b174e3e94031ec1811cce5d535d64bfb619e528ce1075c5"} Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.559513 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.676922 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-inventory-0\") pod \"dc2909c6-f489-467c-a203-01adb27428f5\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.677222 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvdk6\" (UniqueName: \"kubernetes.io/projected/dc2909c6-f489-467c-a203-01adb27428f5-kube-api-access-pvdk6\") pod \"dc2909c6-f489-467c-a203-01adb27428f5\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.677291 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-ssh-key-openstack-edpm-ipam\") pod \"dc2909c6-f489-467c-a203-01adb27428f5\" (UID: \"dc2909c6-f489-467c-a203-01adb27428f5\") " Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.684370 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2909c6-f489-467c-a203-01adb27428f5-kube-api-access-pvdk6" (OuterVolumeSpecName: "kube-api-access-pvdk6") pod "dc2909c6-f489-467c-a203-01adb27428f5" (UID: "dc2909c6-f489-467c-a203-01adb27428f5"). InnerVolumeSpecName "kube-api-access-pvdk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.712068 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dc2909c6-f489-467c-a203-01adb27428f5" (UID: "dc2909c6-f489-467c-a203-01adb27428f5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.714934 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "dc2909c6-f489-467c-a203-01adb27428f5" (UID: "dc2909c6-f489-467c-a203-01adb27428f5"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.780179 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvdk6\" (UniqueName: \"kubernetes.io/projected/dc2909c6-f489-467c-a203-01adb27428f5-kube-api-access-pvdk6\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.780245 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 05 08:49:59 crc kubenswrapper[4876]: I1205 08:49:59.780258 4876 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/dc2909c6-f489-467c-a203-01adb27428f5-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.113483 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" event={"ID":"dc2909c6-f489-467c-a203-01adb27428f5","Type":"ContainerDied","Data":"cddccff364d8a1ada5ca961edcde7e51065605b667703c08d12488bf02b83bd6"} Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.113530 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cddccff364d8a1ada5ca961edcde7e51065605b667703c08d12488bf02b83bd6" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.113596 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-stmmf" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.214918 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987"] Dec 05 08:50:00 crc kubenswrapper[4876]: E1205 08:50:00.215355 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2909c6-f489-467c-a203-01adb27428f5" containerName="ssh-known-hosts-edpm-deployment" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.215374 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2909c6-f489-467c-a203-01adb27428f5" containerName="ssh-known-hosts-edpm-deployment" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.215553 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2909c6-f489-467c-a203-01adb27428f5" containerName="ssh-known-hosts-edpm-deployment" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.216184 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.219600 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.219800 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.219943 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.219997 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.233954 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987"] Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.391410 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.391789 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.392251 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b645r\" (UniqueName: \"kubernetes.io/projected/87e85de2-b888-48c3-8f2d-e500ccedca3e-kube-api-access-b645r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.493823 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.493971 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.494063 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b645r\" (UniqueName: \"kubernetes.io/projected/87e85de2-b888-48c3-8f2d-e500ccedca3e-kube-api-access-b645r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.498702 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.499586 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.512560 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b645r\" (UniqueName: \"kubernetes.io/projected/87e85de2-b888-48c3-8f2d-e500ccedca3e-kube-api-access-b645r\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-8w987\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:00 crc kubenswrapper[4876]: I1205 08:50:00.592513 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:01 crc kubenswrapper[4876]: I1205 08:50:01.149462 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987"] Dec 05 08:50:01 crc kubenswrapper[4876]: I1205 08:50:01.152346 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:50:02 crc kubenswrapper[4876]: I1205 08:50:02.129530 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" event={"ID":"87e85de2-b888-48c3-8f2d-e500ccedca3e","Type":"ContainerStarted","Data":"08bc94042c1c440936a0a36cdf9690b8094f453fe1cb3ca318f68c6873c1096e"} Dec 05 08:50:02 crc kubenswrapper[4876]: I1205 08:50:02.130130 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" event={"ID":"87e85de2-b888-48c3-8f2d-e500ccedca3e","Type":"ContainerStarted","Data":"1980406dd5018bfc7b35453b9836d78ed74e947ba78decead1985978fbef7075"} Dec 05 08:50:02 crc kubenswrapper[4876]: I1205 08:50:02.153389 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" podStartSLOduration=1.69295307 podStartE2EDuration="2.153371584s" podCreationTimestamp="2025-12-05 08:50:00 +0000 UTC" firstStartedPulling="2025-12-05 08:50:01.152154277 +0000 UTC m=+2065.640818899" lastFinishedPulling="2025-12-05 08:50:01.612572791 +0000 UTC m=+2066.101237413" observedRunningTime="2025-12-05 08:50:02.145821628 +0000 UTC m=+2066.634486250" watchObservedRunningTime="2025-12-05 08:50:02.153371584 +0000 UTC m=+2066.642036206" Dec 05 08:50:09 crc kubenswrapper[4876]: E1205 08:50:09.341953 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87e85de2_b888_48c3_8f2d_e500ccedca3e.slice/crio-08bc94042c1c440936a0a36cdf9690b8094f453fe1cb3ca318f68c6873c1096e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87e85de2_b888_48c3_8f2d_e500ccedca3e.slice/crio-conmon-08bc94042c1c440936a0a36cdf9690b8094f453fe1cb3ca318f68c6873c1096e.scope\": RecentStats: unable to find data in memory cache]" Dec 05 08:50:10 crc kubenswrapper[4876]: I1205 08:50:10.201473 4876 generic.go:334] "Generic (PLEG): container finished" podID="87e85de2-b888-48c3-8f2d-e500ccedca3e" containerID="08bc94042c1c440936a0a36cdf9690b8094f453fe1cb3ca318f68c6873c1096e" exitCode=0 Dec 05 08:50:10 crc kubenswrapper[4876]: I1205 08:50:10.201632 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" event={"ID":"87e85de2-b888-48c3-8f2d-e500ccedca3e","Type":"ContainerDied","Data":"08bc94042c1c440936a0a36cdf9690b8094f453fe1cb3ca318f68c6873c1096e"} Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.710106 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.838271 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b645r\" (UniqueName: \"kubernetes.io/projected/87e85de2-b888-48c3-8f2d-e500ccedca3e-kube-api-access-b645r\") pod \"87e85de2-b888-48c3-8f2d-e500ccedca3e\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.838375 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-ssh-key\") pod \"87e85de2-b888-48c3-8f2d-e500ccedca3e\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.838709 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-inventory\") pod \"87e85de2-b888-48c3-8f2d-e500ccedca3e\" (UID: \"87e85de2-b888-48c3-8f2d-e500ccedca3e\") " Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.844038 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e85de2-b888-48c3-8f2d-e500ccedca3e-kube-api-access-b645r" (OuterVolumeSpecName: "kube-api-access-b645r") pod "87e85de2-b888-48c3-8f2d-e500ccedca3e" (UID: "87e85de2-b888-48c3-8f2d-e500ccedca3e"). InnerVolumeSpecName "kube-api-access-b645r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.866034 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "87e85de2-b888-48c3-8f2d-e500ccedca3e" (UID: "87e85de2-b888-48c3-8f2d-e500ccedca3e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.871796 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-inventory" (OuterVolumeSpecName: "inventory") pod "87e85de2-b888-48c3-8f2d-e500ccedca3e" (UID: "87e85de2-b888-48c3-8f2d-e500ccedca3e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.941668 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.941734 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b645r\" (UniqueName: \"kubernetes.io/projected/87e85de2-b888-48c3-8f2d-e500ccedca3e-kube-api-access-b645r\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:11 crc kubenswrapper[4876]: I1205 08:50:11.941752 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87e85de2-b888-48c3-8f2d-e500ccedca3e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.225296 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" event={"ID":"87e85de2-b888-48c3-8f2d-e500ccedca3e","Type":"ContainerDied","Data":"1980406dd5018bfc7b35453b9836d78ed74e947ba78decead1985978fbef7075"} Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.225335 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1980406dd5018bfc7b35453b9836d78ed74e947ba78decead1985978fbef7075" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.225354 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-8w987" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.314958 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w"] Dec 05 08:50:12 crc kubenswrapper[4876]: E1205 08:50:12.315514 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e85de2-b888-48c3-8f2d-e500ccedca3e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.315553 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e85de2-b888-48c3-8f2d-e500ccedca3e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.315984 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e85de2-b888-48c3-8f2d-e500ccedca3e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.316871 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.322312 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.322452 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.322598 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.322656 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.327165 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w"] Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.349375 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.349431 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.349497 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f25z\" (UniqueName: \"kubernetes.io/projected/a4528785-cb5d-4281-9de7-2140359d7c0c-kube-api-access-9f25z\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.451303 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.451617 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.451679 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f25z\" (UniqueName: \"kubernetes.io/projected/a4528785-cb5d-4281-9de7-2140359d7c0c-kube-api-access-9f25z\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.455764 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.464915 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.467880 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f25z\" (UniqueName: \"kubernetes.io/projected/a4528785-cb5d-4281-9de7-2140359d7c0c-kube-api-access-9f25z\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:12 crc kubenswrapper[4876]: I1205 08:50:12.634492 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:13 crc kubenswrapper[4876]: I1205 08:50:13.147025 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w"] Dec 05 08:50:13 crc kubenswrapper[4876]: I1205 08:50:13.234267 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" event={"ID":"a4528785-cb5d-4281-9de7-2140359d7c0c","Type":"ContainerStarted","Data":"7e7446afc2692a7650179d785ab356deb1361905a49aaa02ed426195374798a7"} Dec 05 08:50:14 crc kubenswrapper[4876]: I1205 08:50:14.247891 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" event={"ID":"a4528785-cb5d-4281-9de7-2140359d7c0c","Type":"ContainerStarted","Data":"816992fa5f693232b735fe2b3670db9c7a83c45fd767e917eb948a819ed61148"} Dec 05 08:50:23 crc kubenswrapper[4876]: I1205 08:50:23.321118 4876 generic.go:334] "Generic (PLEG): container finished" podID="a4528785-cb5d-4281-9de7-2140359d7c0c" containerID="816992fa5f693232b735fe2b3670db9c7a83c45fd767e917eb948a819ed61148" exitCode=0 Dec 05 08:50:23 crc kubenswrapper[4876]: I1205 08:50:23.321196 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" event={"ID":"a4528785-cb5d-4281-9de7-2140359d7c0c","Type":"ContainerDied","Data":"816992fa5f693232b735fe2b3670db9c7a83c45fd767e917eb948a819ed61148"} Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.737349 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.807575 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-ssh-key\") pod \"a4528785-cb5d-4281-9de7-2140359d7c0c\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.808385 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-inventory\") pod \"a4528785-cb5d-4281-9de7-2140359d7c0c\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.808632 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f25z\" (UniqueName: \"kubernetes.io/projected/a4528785-cb5d-4281-9de7-2140359d7c0c-kube-api-access-9f25z\") pod \"a4528785-cb5d-4281-9de7-2140359d7c0c\" (UID: \"a4528785-cb5d-4281-9de7-2140359d7c0c\") " Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.819294 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4528785-cb5d-4281-9de7-2140359d7c0c-kube-api-access-9f25z" (OuterVolumeSpecName: "kube-api-access-9f25z") pod "a4528785-cb5d-4281-9de7-2140359d7c0c" (UID: "a4528785-cb5d-4281-9de7-2140359d7c0c"). InnerVolumeSpecName "kube-api-access-9f25z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.839346 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a4528785-cb5d-4281-9de7-2140359d7c0c" (UID: "a4528785-cb5d-4281-9de7-2140359d7c0c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.843038 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-inventory" (OuterVolumeSpecName: "inventory") pod "a4528785-cb5d-4281-9de7-2140359d7c0c" (UID: "a4528785-cb5d-4281-9de7-2140359d7c0c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.911571 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.911831 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a4528785-cb5d-4281-9de7-2140359d7c0c-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:24 crc kubenswrapper[4876]: I1205 08:50:24.911994 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f25z\" (UniqueName: \"kubernetes.io/projected/a4528785-cb5d-4281-9de7-2140359d7c0c-kube-api-access-9f25z\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.338926 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" event={"ID":"a4528785-cb5d-4281-9de7-2140359d7c0c","Type":"ContainerDied","Data":"7e7446afc2692a7650179d785ab356deb1361905a49aaa02ed426195374798a7"} Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.338975 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e7446afc2692a7650179d785ab356deb1361905a49aaa02ed426195374798a7" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.339033 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.405133 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b"] Dec 05 08:50:25 crc kubenswrapper[4876]: E1205 08:50:25.405609 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4528785-cb5d-4281-9de7-2140359d7c0c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.405638 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4528785-cb5d-4281-9de7-2140359d7c0c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.405841 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4528785-cb5d-4281-9de7-2140359d7c0c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.406545 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.412489 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.412569 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.412683 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.412834 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.412848 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.413043 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.413212 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.413370 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.416619 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b"] Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522125 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522194 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522266 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522357 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522389 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522431 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522503 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522622 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522734 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.522808 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.523079 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2rp5\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-kube-api-access-v2rp5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.523115 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.523164 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.523189 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.624890 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625007 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625050 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625084 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625107 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625129 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2rp5\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-kube-api-access-v2rp5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625145 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625169 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625187 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625232 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625255 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625283 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625345 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.625361 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.628865 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.628908 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.630049 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.630452 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.630959 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.631265 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.631577 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.631621 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.632290 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.632631 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.633036 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.634879 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.638770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.643988 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2rp5\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-kube-api-access-v2rp5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:25 crc kubenswrapper[4876]: I1205 08:50:25.726415 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:50:26 crc kubenswrapper[4876]: I1205 08:50:26.241016 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b"] Dec 05 08:50:26 crc kubenswrapper[4876]: I1205 08:50:26.349251 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" event={"ID":"ecc3ee00-cb0c-4084-ba46-80684e18f07e","Type":"ContainerStarted","Data":"ec8fad4afa508eef3c05c3579621801605391ee953df0da66a72aa7ad0fd1255"} Dec 05 08:50:27 crc kubenswrapper[4876]: I1205 08:50:27.357745 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" event={"ID":"ecc3ee00-cb0c-4084-ba46-80684e18f07e","Type":"ContainerStarted","Data":"6bd1b9e075924f2e0832fe2d6ee51f121255ab81c909b496967dd430aee20541"} Dec 05 08:50:27 crc kubenswrapper[4876]: I1205 08:50:27.383714 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" podStartSLOduration=1.982010544 podStartE2EDuration="2.383696744s" podCreationTimestamp="2025-12-05 08:50:25 +0000 UTC" firstStartedPulling="2025-12-05 08:50:26.250968265 +0000 UTC m=+2090.739632887" lastFinishedPulling="2025-12-05 08:50:26.652654465 +0000 UTC m=+2091.141319087" observedRunningTime="2025-12-05 08:50:27.376529722 +0000 UTC m=+2091.865194364" watchObservedRunningTime="2025-12-05 08:50:27.383696744 +0000 UTC m=+2091.872361366" Dec 05 08:50:45 crc kubenswrapper[4876]: I1205 08:50:45.973219 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z2b7p"] Dec 05 08:50:45 crc kubenswrapper[4876]: I1205 08:50:45.975760 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:45 crc kubenswrapper[4876]: I1205 08:50:45.988475 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2b7p"] Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.021393 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hccc\" (UniqueName: \"kubernetes.io/projected/a9a27fc2-a57b-4a00-afaf-eb4628685707-kube-api-access-7hccc\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.021685 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-catalog-content\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.021832 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-utilities\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.123990 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hccc\" (UniqueName: \"kubernetes.io/projected/a9a27fc2-a57b-4a00-afaf-eb4628685707-kube-api-access-7hccc\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.124066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-catalog-content\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.124138 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-utilities\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.124819 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-utilities\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.124923 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-catalog-content\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.142445 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hccc\" (UniqueName: \"kubernetes.io/projected/a9a27fc2-a57b-4a00-afaf-eb4628685707-kube-api-access-7hccc\") pod \"redhat-marketplace-z2b7p\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.297098 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:46 crc kubenswrapper[4876]: I1205 08:50:46.794692 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2b7p"] Dec 05 08:50:46 crc kubenswrapper[4876]: W1205 08:50:46.804668 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a27fc2_a57b_4a00_afaf_eb4628685707.slice/crio-043aaf6d9b0b5a3ab69a612e4cfbc268100570088184e9d51b896f666fe9f830 WatchSource:0}: Error finding container 043aaf6d9b0b5a3ab69a612e4cfbc268100570088184e9d51b896f666fe9f830: Status 404 returned error can't find the container with id 043aaf6d9b0b5a3ab69a612e4cfbc268100570088184e9d51b896f666fe9f830 Dec 05 08:50:47 crc kubenswrapper[4876]: I1205 08:50:47.537426 4876 generic.go:334] "Generic (PLEG): container finished" podID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerID="005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b" exitCode=0 Dec 05 08:50:47 crc kubenswrapper[4876]: I1205 08:50:47.537549 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2b7p" event={"ID":"a9a27fc2-a57b-4a00-afaf-eb4628685707","Type":"ContainerDied","Data":"005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b"} Dec 05 08:50:47 crc kubenswrapper[4876]: I1205 08:50:47.538704 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2b7p" event={"ID":"a9a27fc2-a57b-4a00-afaf-eb4628685707","Type":"ContainerStarted","Data":"043aaf6d9b0b5a3ab69a612e4cfbc268100570088184e9d51b896f666fe9f830"} Dec 05 08:50:48 crc kubenswrapper[4876]: I1205 08:50:48.549221 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2b7p" event={"ID":"a9a27fc2-a57b-4a00-afaf-eb4628685707","Type":"ContainerStarted","Data":"1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7"} Dec 05 08:50:49 crc kubenswrapper[4876]: I1205 08:50:49.559849 4876 generic.go:334] "Generic (PLEG): container finished" podID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerID="1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7" exitCode=0 Dec 05 08:50:49 crc kubenswrapper[4876]: I1205 08:50:49.559888 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2b7p" event={"ID":"a9a27fc2-a57b-4a00-afaf-eb4628685707","Type":"ContainerDied","Data":"1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7"} Dec 05 08:50:50 crc kubenswrapper[4876]: I1205 08:50:50.568602 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2b7p" event={"ID":"a9a27fc2-a57b-4a00-afaf-eb4628685707","Type":"ContainerStarted","Data":"6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c"} Dec 05 08:50:50 crc kubenswrapper[4876]: I1205 08:50:50.588663 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z2b7p" podStartSLOduration=3.128761199 podStartE2EDuration="5.588643876s" podCreationTimestamp="2025-12-05 08:50:45 +0000 UTC" firstStartedPulling="2025-12-05 08:50:47.539290979 +0000 UTC m=+2112.027955601" lastFinishedPulling="2025-12-05 08:50:49.999173616 +0000 UTC m=+2114.487838278" observedRunningTime="2025-12-05 08:50:50.582728698 +0000 UTC m=+2115.071393330" watchObservedRunningTime="2025-12-05 08:50:50.588643876 +0000 UTC m=+2115.077308498" Dec 05 08:50:56 crc kubenswrapper[4876]: I1205 08:50:56.297841 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:56 crc kubenswrapper[4876]: I1205 08:50:56.298486 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:56 crc kubenswrapper[4876]: I1205 08:50:56.343402 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:56 crc kubenswrapper[4876]: I1205 08:50:56.681054 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:56 crc kubenswrapper[4876]: I1205 08:50:56.726502 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2b7p"] Dec 05 08:50:58 crc kubenswrapper[4876]: I1205 08:50:58.639487 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z2b7p" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerName="registry-server" containerID="cri-o://6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c" gracePeriod=2 Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.145355 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.281026 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-catalog-content\") pod \"a9a27fc2-a57b-4a00-afaf-eb4628685707\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.281355 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hccc\" (UniqueName: \"kubernetes.io/projected/a9a27fc2-a57b-4a00-afaf-eb4628685707-kube-api-access-7hccc\") pod \"a9a27fc2-a57b-4a00-afaf-eb4628685707\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.281475 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-utilities\") pod \"a9a27fc2-a57b-4a00-afaf-eb4628685707\" (UID: \"a9a27fc2-a57b-4a00-afaf-eb4628685707\") " Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.282148 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-utilities" (OuterVolumeSpecName: "utilities") pod "a9a27fc2-a57b-4a00-afaf-eb4628685707" (UID: "a9a27fc2-a57b-4a00-afaf-eb4628685707"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.287444 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9a27fc2-a57b-4a00-afaf-eb4628685707-kube-api-access-7hccc" (OuterVolumeSpecName: "kube-api-access-7hccc") pod "a9a27fc2-a57b-4a00-afaf-eb4628685707" (UID: "a9a27fc2-a57b-4a00-afaf-eb4628685707"). InnerVolumeSpecName "kube-api-access-7hccc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.301850 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9a27fc2-a57b-4a00-afaf-eb4628685707" (UID: "a9a27fc2-a57b-4a00-afaf-eb4628685707"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.383417 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.383452 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9a27fc2-a57b-4a00-afaf-eb4628685707-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.383462 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hccc\" (UniqueName: \"kubernetes.io/projected/a9a27fc2-a57b-4a00-afaf-eb4628685707-kube-api-access-7hccc\") on node \"crc\" DevicePath \"\"" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.649954 4876 generic.go:334] "Generic (PLEG): container finished" podID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerID="6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c" exitCode=0 Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.649999 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2b7p" event={"ID":"a9a27fc2-a57b-4a00-afaf-eb4628685707","Type":"ContainerDied","Data":"6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c"} Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.650022 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2b7p" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.650040 4876 scope.go:117] "RemoveContainer" containerID="6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.650029 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2b7p" event={"ID":"a9a27fc2-a57b-4a00-afaf-eb4628685707","Type":"ContainerDied","Data":"043aaf6d9b0b5a3ab69a612e4cfbc268100570088184e9d51b896f666fe9f830"} Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.679840 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2b7p"] Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.684353 4876 scope.go:117] "RemoveContainer" containerID="1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.694177 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2b7p"] Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.707725 4876 scope.go:117] "RemoveContainer" containerID="005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.743476 4876 scope.go:117] "RemoveContainer" containerID="6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c" Dec 05 08:50:59 crc kubenswrapper[4876]: E1205 08:50:59.744108 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c\": container with ID starting with 6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c not found: ID does not exist" containerID="6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.744165 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c"} err="failed to get container status \"6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c\": rpc error: code = NotFound desc = could not find container \"6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c\": container with ID starting with 6edfa51c656a19b8713493fe7d2668b4ed1c6c02030eaa270ecffd861bfc2e4c not found: ID does not exist" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.744197 4876 scope.go:117] "RemoveContainer" containerID="1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7" Dec 05 08:50:59 crc kubenswrapper[4876]: E1205 08:50:59.744650 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7\": container with ID starting with 1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7 not found: ID does not exist" containerID="1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.744698 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7"} err="failed to get container status \"1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7\": rpc error: code = NotFound desc = could not find container \"1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7\": container with ID starting with 1d2035f353d1f6d87e94c188657527799c7edf772bda418c72daf7a0c328e8f7 not found: ID does not exist" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.744725 4876 scope.go:117] "RemoveContainer" containerID="005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b" Dec 05 08:50:59 crc kubenswrapper[4876]: E1205 08:50:59.745122 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b\": container with ID starting with 005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b not found: ID does not exist" containerID="005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.745153 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b"} err="failed to get container status \"005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b\": rpc error: code = NotFound desc = could not find container \"005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b\": container with ID starting with 005e4105a68ae1d16c7d6b8557b878915a3876595fe2c68250a793510303d68b not found: ID does not exist" Dec 05 08:50:59 crc kubenswrapper[4876]: I1205 08:50:59.846819 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" path="/var/lib/kubelet/pods/a9a27fc2-a57b-4a00-afaf-eb4628685707/volumes" Dec 05 08:51:02 crc kubenswrapper[4876]: I1205 08:51:02.690311 4876 generic.go:334] "Generic (PLEG): container finished" podID="ecc3ee00-cb0c-4084-ba46-80684e18f07e" containerID="6bd1b9e075924f2e0832fe2d6ee51f121255ab81c909b496967dd430aee20541" exitCode=0 Dec 05 08:51:02 crc kubenswrapper[4876]: I1205 08:51:02.690492 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" event={"ID":"ecc3ee00-cb0c-4084-ba46-80684e18f07e","Type":"ContainerDied","Data":"6bd1b9e075924f2e0832fe2d6ee51f121255ab81c909b496967dd430aee20541"} Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.080068 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278367 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-ovn-default-certs-0\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278415 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278444 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278510 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-repo-setup-combined-ca-bundle\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278535 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-telemetry-combined-ca-bundle\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278572 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-libvirt-combined-ca-bundle\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278620 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-neutron-metadata-combined-ca-bundle\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278641 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-inventory\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278729 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278777 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ssh-key\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278818 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ovn-combined-ca-bundle\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278835 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-bootstrap-combined-ca-bundle\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278854 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2rp5\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-kube-api-access-v2rp5\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.278879 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-nova-combined-ca-bundle\") pod \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\" (UID: \"ecc3ee00-cb0c-4084-ba46-80684e18f07e\") " Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.287837 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.288444 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.288556 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.288582 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.289148 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.289369 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.289692 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.289866 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.291587 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.291633 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.295348 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.295394 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-kube-api-access-v2rp5" (OuterVolumeSpecName: "kube-api-access-v2rp5") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "kube-api-access-v2rp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.321410 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.329128 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-inventory" (OuterVolumeSpecName: "inventory") pod "ecc3ee00-cb0c-4084-ba46-80684e18f07e" (UID: "ecc3ee00-cb0c-4084-ba46-80684e18f07e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381447 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381505 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381526 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381546 4876 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381564 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2rp5\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-kube-api-access-v2rp5\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381583 4876 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381602 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381621 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381641 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/ecc3ee00-cb0c-4084-ba46-80684e18f07e-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381659 4876 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381678 4876 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381696 4876 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381712 4876 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.381730 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ecc3ee00-cb0c-4084-ba46-80684e18f07e-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.717114 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" event={"ID":"ecc3ee00-cb0c-4084-ba46-80684e18f07e","Type":"ContainerDied","Data":"ec8fad4afa508eef3c05c3579621801605391ee953df0da66a72aa7ad0fd1255"} Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.717157 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec8fad4afa508eef3c05c3579621801605391ee953df0da66a72aa7ad0fd1255" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.717224 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.834836 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k"] Dec 05 08:51:04 crc kubenswrapper[4876]: E1205 08:51:04.836012 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerName="registry-server" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.836126 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerName="registry-server" Dec 05 08:51:04 crc kubenswrapper[4876]: E1205 08:51:04.836227 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerName="extract-utilities" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.836313 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerName="extract-utilities" Dec 05 08:51:04 crc kubenswrapper[4876]: E1205 08:51:04.836420 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc3ee00-cb0c-4084-ba46-80684e18f07e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.836497 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc3ee00-cb0c-4084-ba46-80684e18f07e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 05 08:51:04 crc kubenswrapper[4876]: E1205 08:51:04.836590 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerName="extract-content" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.836664 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerName="extract-content" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.836992 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc3ee00-cb0c-4084-ba46-80684e18f07e" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.837102 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9a27fc2-a57b-4a00-afaf-eb4628685707" containerName="registry-server" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.837851 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.840395 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.840710 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.841654 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.841830 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.844606 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.858259 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k"] Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.891144 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/33523fba-8378-49ee-b023-0f80a4a8f91b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.891225 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.891327 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.891380 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r99gh\" (UniqueName: \"kubernetes.io/projected/33523fba-8378-49ee-b023-0f80a4a8f91b-kube-api-access-r99gh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.891433 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.992765 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r99gh\" (UniqueName: \"kubernetes.io/projected/33523fba-8378-49ee-b023-0f80a4a8f91b-kube-api-access-r99gh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.992826 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.992949 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/33523fba-8378-49ee-b023-0f80a4a8f91b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.992998 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.993081 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.994305 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/33523fba-8378-49ee-b023-0f80a4a8f91b-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:04 crc kubenswrapper[4876]: I1205 08:51:04.999214 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:05 crc kubenswrapper[4876]: I1205 08:51:05.001051 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:05 crc kubenswrapper[4876]: I1205 08:51:05.001292 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:05 crc kubenswrapper[4876]: I1205 08:51:05.013472 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r99gh\" (UniqueName: \"kubernetes.io/projected/33523fba-8378-49ee-b023-0f80a4a8f91b-kube-api-access-r99gh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pdx5k\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:05 crc kubenswrapper[4876]: I1205 08:51:05.161046 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:51:05 crc kubenswrapper[4876]: I1205 08:51:05.702399 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k"] Dec 05 08:51:05 crc kubenswrapper[4876]: W1205 08:51:05.709451 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33523fba_8378_49ee_b023_0f80a4a8f91b.slice/crio-dbde265b8ab584888e2ac7d3103bf923f9a04c083ba2423f06cd0a665c66bfd6 WatchSource:0}: Error finding container dbde265b8ab584888e2ac7d3103bf923f9a04c083ba2423f06cd0a665c66bfd6: Status 404 returned error can't find the container with id dbde265b8ab584888e2ac7d3103bf923f9a04c083ba2423f06cd0a665c66bfd6 Dec 05 08:51:05 crc kubenswrapper[4876]: I1205 08:51:05.726521 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" event={"ID":"33523fba-8378-49ee-b023-0f80a4a8f91b","Type":"ContainerStarted","Data":"dbde265b8ab584888e2ac7d3103bf923f9a04c083ba2423f06cd0a665c66bfd6"} Dec 05 08:51:06 crc kubenswrapper[4876]: I1205 08:51:06.739529 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" event={"ID":"33523fba-8378-49ee-b023-0f80a4a8f91b","Type":"ContainerStarted","Data":"563ef44f1826a7f39be80aae19aec1ebc8a725bee3d5b751af5734b9e72ad31e"} Dec 05 08:51:06 crc kubenswrapper[4876]: I1205 08:51:06.758814 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" podStartSLOduration=2.28665554 podStartE2EDuration="2.758792281s" podCreationTimestamp="2025-12-05 08:51:04 +0000 UTC" firstStartedPulling="2025-12-05 08:51:05.711657139 +0000 UTC m=+2130.200321761" lastFinishedPulling="2025-12-05 08:51:06.18379388 +0000 UTC m=+2130.672458502" observedRunningTime="2025-12-05 08:51:06.757047964 +0000 UTC m=+2131.245712596" watchObservedRunningTime="2025-12-05 08:51:06.758792281 +0000 UTC m=+2131.247456903" Dec 05 08:51:38 crc kubenswrapper[4876]: I1205 08:51:38.214039 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:51:38 crc kubenswrapper[4876]: I1205 08:51:38.214654 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.401136 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7lvqd"] Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.403794 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.412878 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7lvqd"] Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.593913 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-utilities\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.594083 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-catalog-content\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.594155 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-262l9\" (UniqueName: \"kubernetes.io/projected/e9a4126e-fddf-408e-b28c-7aecbc787999-kube-api-access-262l9\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.695726 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-262l9\" (UniqueName: \"kubernetes.io/projected/e9a4126e-fddf-408e-b28c-7aecbc787999-kube-api-access-262l9\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.695834 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-utilities\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.695951 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-catalog-content\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.696408 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-catalog-content\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.696639 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-utilities\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.723970 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-262l9\" (UniqueName: \"kubernetes.io/projected/e9a4126e-fddf-408e-b28c-7aecbc787999-kube-api-access-262l9\") pod \"certified-operators-7lvqd\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:54 crc kubenswrapper[4876]: I1205 08:51:54.730124 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:51:55 crc kubenswrapper[4876]: I1205 08:51:55.288248 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7lvqd"] Dec 05 08:51:56 crc kubenswrapper[4876]: I1205 08:51:56.227094 4876 generic.go:334] "Generic (PLEG): container finished" podID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerID="d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f" exitCode=0 Dec 05 08:51:56 crc kubenswrapper[4876]: I1205 08:51:56.227251 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lvqd" event={"ID":"e9a4126e-fddf-408e-b28c-7aecbc787999","Type":"ContainerDied","Data":"d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f"} Dec 05 08:51:56 crc kubenswrapper[4876]: I1205 08:51:56.227477 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lvqd" event={"ID":"e9a4126e-fddf-408e-b28c-7aecbc787999","Type":"ContainerStarted","Data":"28e2b7049b77978ff1238f0f12320f1be11640e713c01d371d2df4dd5f0d940b"} Dec 05 08:51:58 crc kubenswrapper[4876]: I1205 08:51:58.248398 4876 generic.go:334] "Generic (PLEG): container finished" podID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerID="32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d" exitCode=0 Dec 05 08:51:58 crc kubenswrapper[4876]: I1205 08:51:58.248458 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lvqd" event={"ID":"e9a4126e-fddf-408e-b28c-7aecbc787999","Type":"ContainerDied","Data":"32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d"} Dec 05 08:51:59 crc kubenswrapper[4876]: I1205 08:51:59.260561 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lvqd" event={"ID":"e9a4126e-fddf-408e-b28c-7aecbc787999","Type":"ContainerStarted","Data":"19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7"} Dec 05 08:51:59 crc kubenswrapper[4876]: I1205 08:51:59.284579 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7lvqd" podStartSLOduration=2.852973831 podStartE2EDuration="5.284563829s" podCreationTimestamp="2025-12-05 08:51:54 +0000 UTC" firstStartedPulling="2025-12-05 08:51:56.229067918 +0000 UTC m=+2180.717732560" lastFinishedPulling="2025-12-05 08:51:58.660657936 +0000 UTC m=+2183.149322558" observedRunningTime="2025-12-05 08:51:59.278873326 +0000 UTC m=+2183.767537988" watchObservedRunningTime="2025-12-05 08:51:59.284563829 +0000 UTC m=+2183.773228451" Dec 05 08:52:04 crc kubenswrapper[4876]: I1205 08:52:04.730784 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:52:04 crc kubenswrapper[4876]: I1205 08:52:04.731372 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:52:04 crc kubenswrapper[4876]: I1205 08:52:04.807549 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:52:05 crc kubenswrapper[4876]: I1205 08:52:05.377351 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:52:05 crc kubenswrapper[4876]: I1205 08:52:05.440757 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7lvqd"] Dec 05 08:52:06 crc kubenswrapper[4876]: I1205 08:52:06.337419 4876 generic.go:334] "Generic (PLEG): container finished" podID="33523fba-8378-49ee-b023-0f80a4a8f91b" containerID="563ef44f1826a7f39be80aae19aec1ebc8a725bee3d5b751af5734b9e72ad31e" exitCode=0 Dec 05 08:52:06 crc kubenswrapper[4876]: I1205 08:52:06.337493 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" event={"ID":"33523fba-8378-49ee-b023-0f80a4a8f91b","Type":"ContainerDied","Data":"563ef44f1826a7f39be80aae19aec1ebc8a725bee3d5b751af5734b9e72ad31e"} Dec 05 08:52:07 crc kubenswrapper[4876]: I1205 08:52:07.348123 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7lvqd" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerName="registry-server" containerID="cri-o://19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7" gracePeriod=2 Dec 05 08:52:07 crc kubenswrapper[4876]: I1205 08:52:07.879062 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:52:07 crc kubenswrapper[4876]: I1205 08:52:07.889512 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.046543 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-catalog-content\") pod \"e9a4126e-fddf-408e-b28c-7aecbc787999\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.046709 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ovn-combined-ca-bundle\") pod \"33523fba-8378-49ee-b023-0f80a4a8f91b\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.046760 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r99gh\" (UniqueName: \"kubernetes.io/projected/33523fba-8378-49ee-b023-0f80a4a8f91b-kube-api-access-r99gh\") pod \"33523fba-8378-49ee-b023-0f80a4a8f91b\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.046808 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-utilities\") pod \"e9a4126e-fddf-408e-b28c-7aecbc787999\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.046849 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ssh-key\") pod \"33523fba-8378-49ee-b023-0f80a4a8f91b\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.046974 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-inventory\") pod \"33523fba-8378-49ee-b023-0f80a4a8f91b\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.047012 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-262l9\" (UniqueName: \"kubernetes.io/projected/e9a4126e-fddf-408e-b28c-7aecbc787999-kube-api-access-262l9\") pod \"e9a4126e-fddf-408e-b28c-7aecbc787999\" (UID: \"e9a4126e-fddf-408e-b28c-7aecbc787999\") " Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.047040 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/33523fba-8378-49ee-b023-0f80a4a8f91b-ovncontroller-config-0\") pod \"33523fba-8378-49ee-b023-0f80a4a8f91b\" (UID: \"33523fba-8378-49ee-b023-0f80a4a8f91b\") " Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.047536 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-utilities" (OuterVolumeSpecName: "utilities") pod "e9a4126e-fddf-408e-b28c-7aecbc787999" (UID: "e9a4126e-fddf-408e-b28c-7aecbc787999"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.053279 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33523fba-8378-49ee-b023-0f80a4a8f91b-kube-api-access-r99gh" (OuterVolumeSpecName: "kube-api-access-r99gh") pod "33523fba-8378-49ee-b023-0f80a4a8f91b" (UID: "33523fba-8378-49ee-b023-0f80a4a8f91b"). InnerVolumeSpecName "kube-api-access-r99gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.053733 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "33523fba-8378-49ee-b023-0f80a4a8f91b" (UID: "33523fba-8378-49ee-b023-0f80a4a8f91b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.057332 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a4126e-fddf-408e-b28c-7aecbc787999-kube-api-access-262l9" (OuterVolumeSpecName: "kube-api-access-262l9") pod "e9a4126e-fddf-408e-b28c-7aecbc787999" (UID: "e9a4126e-fddf-408e-b28c-7aecbc787999"). InnerVolumeSpecName "kube-api-access-262l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.075282 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33523fba-8378-49ee-b023-0f80a4a8f91b-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "33523fba-8378-49ee-b023-0f80a4a8f91b" (UID: "33523fba-8378-49ee-b023-0f80a4a8f91b"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.080137 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "33523fba-8378-49ee-b023-0f80a4a8f91b" (UID: "33523fba-8378-49ee-b023-0f80a4a8f91b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.089896 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-inventory" (OuterVolumeSpecName: "inventory") pod "33523fba-8378-49ee-b023-0f80a4a8f91b" (UID: "33523fba-8378-49ee-b023-0f80a4a8f91b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.098472 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9a4126e-fddf-408e-b28c-7aecbc787999" (UID: "e9a4126e-fddf-408e-b28c-7aecbc787999"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.148841 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.148880 4876 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.148893 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r99gh\" (UniqueName: \"kubernetes.io/projected/33523fba-8378-49ee-b023-0f80a4a8f91b-kube-api-access-r99gh\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.148922 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9a4126e-fddf-408e-b28c-7aecbc787999-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.148930 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.148942 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33523fba-8378-49ee-b023-0f80a4a8f91b-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.148950 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-262l9\" (UniqueName: \"kubernetes.io/projected/e9a4126e-fddf-408e-b28c-7aecbc787999-kube-api-access-262l9\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.148959 4876 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/33523fba-8378-49ee-b023-0f80a4a8f91b-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.213642 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.213727 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.361055 4876 generic.go:334] "Generic (PLEG): container finished" podID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerID="19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7" exitCode=0 Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.361119 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lvqd" event={"ID":"e9a4126e-fddf-408e-b28c-7aecbc787999","Type":"ContainerDied","Data":"19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7"} Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.361146 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7lvqd" event={"ID":"e9a4126e-fddf-408e-b28c-7aecbc787999","Type":"ContainerDied","Data":"28e2b7049b77978ff1238f0f12320f1be11640e713c01d371d2df4dd5f0d940b"} Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.361146 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7lvqd" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.361162 4876 scope.go:117] "RemoveContainer" containerID="19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.363964 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" event={"ID":"33523fba-8378-49ee-b023-0f80a4a8f91b","Type":"ContainerDied","Data":"dbde265b8ab584888e2ac7d3103bf923f9a04c083ba2423f06cd0a665c66bfd6"} Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.363989 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbde265b8ab584888e2ac7d3103bf923f9a04c083ba2423f06cd0a665c66bfd6" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.364044 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pdx5k" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.396938 4876 scope.go:117] "RemoveContainer" containerID="32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.427580 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7lvqd"] Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.433555 4876 scope.go:117] "RemoveContainer" containerID="d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.444599 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7lvqd"] Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.453968 4876 scope.go:117] "RemoveContainer" containerID="19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7" Dec 05 08:52:08 crc kubenswrapper[4876]: E1205 08:52:08.454469 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7\": container with ID starting with 19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7 not found: ID does not exist" containerID="19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.454521 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7"} err="failed to get container status \"19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7\": rpc error: code = NotFound desc = could not find container \"19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7\": container with ID starting with 19573f94b150e0a0c6bebc3b788eb1c74a645d15835e0817d3ecbf2ca78d3ce7 not found: ID does not exist" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.454581 4876 scope.go:117] "RemoveContainer" containerID="32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d" Dec 05 08:52:08 crc kubenswrapper[4876]: E1205 08:52:08.462785 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d\": container with ID starting with 32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d not found: ID does not exist" containerID="32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.462833 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d"} err="failed to get container status \"32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d\": rpc error: code = NotFound desc = could not find container \"32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d\": container with ID starting with 32e1e74417fae16cff163146f9b832b3536977e79c5244a02f917999fe53183d not found: ID does not exist" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.462861 4876 scope.go:117] "RemoveContainer" containerID="d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f" Dec 05 08:52:08 crc kubenswrapper[4876]: E1205 08:52:08.463332 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f\": container with ID starting with d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f not found: ID does not exist" containerID="d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.463359 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f"} err="failed to get container status \"d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f\": rpc error: code = NotFound desc = could not find container \"d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f\": container with ID starting with d545b40748884003416a436f66a2653d42c9205464134f7e01e218cda6d3ac6f not found: ID does not exist" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.512951 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c"] Dec 05 08:52:08 crc kubenswrapper[4876]: E1205 08:52:08.513718 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerName="extract-content" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.513743 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerName="extract-content" Dec 05 08:52:08 crc kubenswrapper[4876]: E1205 08:52:08.513759 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerName="extract-utilities" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.513768 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerName="extract-utilities" Dec 05 08:52:08 crc kubenswrapper[4876]: E1205 08:52:08.513787 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33523fba-8378-49ee-b023-0f80a4a8f91b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.513795 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="33523fba-8378-49ee-b023-0f80a4a8f91b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 05 08:52:08 crc kubenswrapper[4876]: E1205 08:52:08.513829 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerName="registry-server" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.513836 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerName="registry-server" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.514081 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" containerName="registry-server" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.514113 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="33523fba-8378-49ee-b023-0f80a4a8f91b" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.514884 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.517439 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.518543 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.519243 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.519383 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.519518 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.519684 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.530002 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c"] Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.662186 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.662251 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pjmj\" (UniqueName: \"kubernetes.io/projected/74b443c1-b03a-42d2-8a66-49510fdfb5bb-kube-api-access-2pjmj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.662283 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.662313 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.662354 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.662425 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.764654 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pjmj\" (UniqueName: \"kubernetes.io/projected/74b443c1-b03a-42d2-8a66-49510fdfb5bb-kube-api-access-2pjmj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.764711 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.764741 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.764775 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.764837 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.765006 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.770763 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.770763 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.772729 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.778156 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.779400 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.781677 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pjmj\" (UniqueName: \"kubernetes.io/projected/74b443c1-b03a-42d2-8a66-49510fdfb5bb-kube-api-access-2pjmj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:08 crc kubenswrapper[4876]: I1205 08:52:08.851070 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:09 crc kubenswrapper[4876]: I1205 08:52:09.412283 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c"] Dec 05 08:52:09 crc kubenswrapper[4876]: I1205 08:52:09.848399 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9a4126e-fddf-408e-b28c-7aecbc787999" path="/var/lib/kubelet/pods/e9a4126e-fddf-408e-b28c-7aecbc787999/volumes" Dec 05 08:52:10 crc kubenswrapper[4876]: I1205 08:52:10.381547 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" event={"ID":"74b443c1-b03a-42d2-8a66-49510fdfb5bb","Type":"ContainerStarted","Data":"88c86b7d8070a9ce021550482ee1549ba1b531dc09563aeb462cc21ce196e555"} Dec 05 08:52:10 crc kubenswrapper[4876]: I1205 08:52:10.381605 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" event={"ID":"74b443c1-b03a-42d2-8a66-49510fdfb5bb","Type":"ContainerStarted","Data":"304404923ddf3a6e496ad97b6ef8c78014c1d8c10d9b78e045ce3179c1426a59"} Dec 05 08:52:11 crc kubenswrapper[4876]: I1205 08:52:11.416683 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" podStartSLOduration=3.012754005 podStartE2EDuration="3.416664324s" podCreationTimestamp="2025-12-05 08:52:08 +0000 UTC" firstStartedPulling="2025-12-05 08:52:09.412175049 +0000 UTC m=+2193.900839671" lastFinishedPulling="2025-12-05 08:52:09.816085368 +0000 UTC m=+2194.304749990" observedRunningTime="2025-12-05 08:52:11.407788876 +0000 UTC m=+2195.896453498" watchObservedRunningTime="2025-12-05 08:52:11.416664324 +0000 UTC m=+2195.905328946" Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.213555 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.214294 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.214881 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.215922 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.216022 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" gracePeriod=600 Dec 05 08:52:38 crc kubenswrapper[4876]: E1205 08:52:38.340154 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.646920 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" exitCode=0 Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.646959 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412"} Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.647018 4876 scope.go:117] "RemoveContainer" containerID="5ca3218eccf9f7164d33de275ebcdaf5bcf4922b876f83c58de4aed55797c0a8" Dec 05 08:52:38 crc kubenswrapper[4876]: I1205 08:52:38.647731 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:52:38 crc kubenswrapper[4876]: E1205 08:52:38.648103 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:52:53 crc kubenswrapper[4876]: I1205 08:52:53.824870 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:52:53 crc kubenswrapper[4876]: E1205 08:52:53.828626 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:52:56 crc kubenswrapper[4876]: I1205 08:52:56.868955 4876 generic.go:334] "Generic (PLEG): container finished" podID="74b443c1-b03a-42d2-8a66-49510fdfb5bb" containerID="88c86b7d8070a9ce021550482ee1549ba1b531dc09563aeb462cc21ce196e555" exitCode=0 Dec 05 08:52:56 crc kubenswrapper[4876]: I1205 08:52:56.869057 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" event={"ID":"74b443c1-b03a-42d2-8a66-49510fdfb5bb","Type":"ContainerDied","Data":"88c86b7d8070a9ce021550482ee1549ba1b531dc09563aeb462cc21ce196e555"} Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.329798 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.401911 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-ssh-key\") pod \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.402030 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pjmj\" (UniqueName: \"kubernetes.io/projected/74b443c1-b03a-42d2-8a66-49510fdfb5bb-kube-api-access-2pjmj\") pod \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.402776 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.402827 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-metadata-combined-ca-bundle\") pod \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.402875 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-nova-metadata-neutron-config-0\") pod \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.402956 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory\") pod \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.411147 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "74b443c1-b03a-42d2-8a66-49510fdfb5bb" (UID: "74b443c1-b03a-42d2-8a66-49510fdfb5bb"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.422638 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b443c1-b03a-42d2-8a66-49510fdfb5bb-kube-api-access-2pjmj" (OuterVolumeSpecName: "kube-api-access-2pjmj") pod "74b443c1-b03a-42d2-8a66-49510fdfb5bb" (UID: "74b443c1-b03a-42d2-8a66-49510fdfb5bb"). InnerVolumeSpecName "kube-api-access-2pjmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.442135 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "74b443c1-b03a-42d2-8a66-49510fdfb5bb" (UID: "74b443c1-b03a-42d2-8a66-49510fdfb5bb"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.450587 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74b443c1-b03a-42d2-8a66-49510fdfb5bb" (UID: "74b443c1-b03a-42d2-8a66-49510fdfb5bb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:52:58 crc kubenswrapper[4876]: E1205 08:52:58.450700 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory podName:74b443c1-b03a-42d2-8a66-49510fdfb5bb nodeName:}" failed. No retries permitted until 2025-12-05 08:52:58.950668993 +0000 UTC m=+2243.439333615 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory") pod "74b443c1-b03a-42d2-8a66-49510fdfb5bb" (UID: "74b443c1-b03a-42d2-8a66-49510fdfb5bb") : error deleting /var/lib/kubelet/pods/74b443c1-b03a-42d2-8a66-49510fdfb5bb/volume-subpaths: remove /var/lib/kubelet/pods/74b443c1-b03a-42d2-8a66-49510fdfb5bb/volume-subpaths: no such file or directory Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.454056 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "74b443c1-b03a-42d2-8a66-49510fdfb5bb" (UID: "74b443c1-b03a-42d2-8a66-49510fdfb5bb"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.505577 4876 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.505617 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.505628 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pjmj\" (UniqueName: \"kubernetes.io/projected/74b443c1-b03a-42d2-8a66-49510fdfb5bb-kube-api-access-2pjmj\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.505638 4876 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.505647 4876 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.895607 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" event={"ID":"74b443c1-b03a-42d2-8a66-49510fdfb5bb","Type":"ContainerDied","Data":"304404923ddf3a6e496ad97b6ef8c78014c1d8c10d9b78e045ce3179c1426a59"} Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.895845 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="304404923ddf3a6e496ad97b6ef8c78014c1d8c10d9b78e045ce3179c1426a59" Dec 05 08:52:58 crc kubenswrapper[4876]: I1205 08:52:58.895664 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.018372 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory\") pod \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\" (UID: \"74b443c1-b03a-42d2-8a66-49510fdfb5bb\") " Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.034041 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl"] Dec 05 08:52:59 crc kubenswrapper[4876]: E1205 08:52:59.035249 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b443c1-b03a-42d2-8a66-49510fdfb5bb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.035292 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b443c1-b03a-42d2-8a66-49510fdfb5bb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.035949 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b443c1-b03a-42d2-8a66-49510fdfb5bb" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.040574 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.041106 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory" (OuterVolumeSpecName: "inventory") pod "74b443c1-b03a-42d2-8a66-49510fdfb5bb" (UID: "74b443c1-b03a-42d2-8a66-49510fdfb5bb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.043861 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.063317 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl"] Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.121198 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.123260 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.123572 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.123667 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mhn5\" (UniqueName: \"kubernetes.io/projected/21e86dc4-3382-464d-aa9b-a811ee65e5d1-kube-api-access-7mhn5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.123742 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.123974 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74b443c1-b03a-42d2-8a66-49510fdfb5bb-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.225863 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.225952 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mhn5\" (UniqueName: \"kubernetes.io/projected/21e86dc4-3382-464d-aa9b-a811ee65e5d1-kube-api-access-7mhn5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.225983 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.226035 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.226077 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.229438 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.229534 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.230052 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.230535 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.243131 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mhn5\" (UniqueName: \"kubernetes.io/projected/21e86dc4-3382-464d-aa9b-a811ee65e5d1-kube-api-access-7mhn5\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:52:59 crc kubenswrapper[4876]: I1205 08:52:59.397294 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:53:00 crc kubenswrapper[4876]: I1205 08:53:00.016390 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl"] Dec 05 08:53:00 crc kubenswrapper[4876]: I1205 08:53:00.916065 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" event={"ID":"21e86dc4-3382-464d-aa9b-a811ee65e5d1","Type":"ContainerStarted","Data":"fda7a2fd8cfe8f1b0c41f7e418852679ca3fcde3737c7766cde89d73fc1b4097"} Dec 05 08:53:00 crc kubenswrapper[4876]: I1205 08:53:00.916372 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" event={"ID":"21e86dc4-3382-464d-aa9b-a811ee65e5d1","Type":"ContainerStarted","Data":"8867f09788afd50df8feeabe181267d592fbe32296df7f1c6acb6ba4e2183687"} Dec 05 08:53:00 crc kubenswrapper[4876]: I1205 08:53:00.942494 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" podStartSLOduration=2.406853081 podStartE2EDuration="2.942474445s" podCreationTimestamp="2025-12-05 08:52:58 +0000 UTC" firstStartedPulling="2025-12-05 08:53:00.021759615 +0000 UTC m=+2244.510424237" lastFinishedPulling="2025-12-05 08:53:00.557380959 +0000 UTC m=+2245.046045601" observedRunningTime="2025-12-05 08:53:00.929263458 +0000 UTC m=+2245.417928080" watchObservedRunningTime="2025-12-05 08:53:00.942474445 +0000 UTC m=+2245.431139057" Dec 05 08:53:08 crc kubenswrapper[4876]: I1205 08:53:08.823917 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:53:08 crc kubenswrapper[4876]: E1205 08:53:08.824651 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:53:23 crc kubenswrapper[4876]: I1205 08:53:23.824196 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:53:23 crc kubenswrapper[4876]: E1205 08:53:23.825696 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:53:37 crc kubenswrapper[4876]: I1205 08:53:37.825104 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:53:37 crc kubenswrapper[4876]: E1205 08:53:37.826076 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:53:52 crc kubenswrapper[4876]: I1205 08:53:52.824435 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:53:52 crc kubenswrapper[4876]: E1205 08:53:52.825314 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:54:06 crc kubenswrapper[4876]: I1205 08:54:06.825244 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:54:06 crc kubenswrapper[4876]: E1205 08:54:06.825885 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:54:18 crc kubenswrapper[4876]: I1205 08:54:18.824405 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:54:18 crc kubenswrapper[4876]: E1205 08:54:18.825419 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:54:31 crc kubenswrapper[4876]: I1205 08:54:31.824472 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:54:31 crc kubenswrapper[4876]: E1205 08:54:31.825397 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:54:45 crc kubenswrapper[4876]: I1205 08:54:45.839835 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:54:45 crc kubenswrapper[4876]: E1205 08:54:45.840753 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:54:59 crc kubenswrapper[4876]: I1205 08:54:59.823447 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:54:59 crc kubenswrapper[4876]: E1205 08:54:59.824078 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:55:14 crc kubenswrapper[4876]: I1205 08:55:14.823953 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:55:14 crc kubenswrapper[4876]: E1205 08:55:14.824736 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:55:29 crc kubenswrapper[4876]: I1205 08:55:29.824426 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:55:29 crc kubenswrapper[4876]: E1205 08:55:29.825198 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:55:44 crc kubenswrapper[4876]: I1205 08:55:44.824358 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:55:44 crc kubenswrapper[4876]: E1205 08:55:44.825382 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:55:57 crc kubenswrapper[4876]: I1205 08:55:57.824576 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:55:57 crc kubenswrapper[4876]: E1205 08:55:57.825531 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:56:12 crc kubenswrapper[4876]: I1205 08:56:12.824184 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:56:12 crc kubenswrapper[4876]: E1205 08:56:12.824873 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:56:27 crc kubenswrapper[4876]: I1205 08:56:27.825062 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:56:27 crc kubenswrapper[4876]: E1205 08:56:27.826012 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:56:39 crc kubenswrapper[4876]: I1205 08:56:39.824288 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:56:39 crc kubenswrapper[4876]: E1205 08:56:39.825139 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:56:54 crc kubenswrapper[4876]: I1205 08:56:54.824707 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:56:54 crc kubenswrapper[4876]: E1205 08:56:54.825487 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:57:06 crc kubenswrapper[4876]: I1205 08:57:06.824181 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:57:06 crc kubenswrapper[4876]: E1205 08:57:06.824967 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:57:10 crc kubenswrapper[4876]: I1205 08:57:10.256666 4876 generic.go:334] "Generic (PLEG): container finished" podID="21e86dc4-3382-464d-aa9b-a811ee65e5d1" containerID="fda7a2fd8cfe8f1b0c41f7e418852679ca3fcde3737c7766cde89d73fc1b4097" exitCode=0 Dec 05 08:57:10 crc kubenswrapper[4876]: I1205 08:57:10.256748 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" event={"ID":"21e86dc4-3382-464d-aa9b-a811ee65e5d1","Type":"ContainerDied","Data":"fda7a2fd8cfe8f1b0c41f7e418852679ca3fcde3737c7766cde89d73fc1b4097"} Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.669660 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.870247 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-inventory\") pod \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.870336 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mhn5\" (UniqueName: \"kubernetes.io/projected/21e86dc4-3382-464d-aa9b-a811ee65e5d1-kube-api-access-7mhn5\") pod \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.870401 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-ssh-key\") pod \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.870435 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-secret-0\") pod \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.870577 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-combined-ca-bundle\") pod \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\" (UID: \"21e86dc4-3382-464d-aa9b-a811ee65e5d1\") " Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.876658 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "21e86dc4-3382-464d-aa9b-a811ee65e5d1" (UID: "21e86dc4-3382-464d-aa9b-a811ee65e5d1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.877057 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e86dc4-3382-464d-aa9b-a811ee65e5d1-kube-api-access-7mhn5" (OuterVolumeSpecName: "kube-api-access-7mhn5") pod "21e86dc4-3382-464d-aa9b-a811ee65e5d1" (UID: "21e86dc4-3382-464d-aa9b-a811ee65e5d1"). InnerVolumeSpecName "kube-api-access-7mhn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.903660 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "21e86dc4-3382-464d-aa9b-a811ee65e5d1" (UID: "21e86dc4-3382-464d-aa9b-a811ee65e5d1"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.904225 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-inventory" (OuterVolumeSpecName: "inventory") pod "21e86dc4-3382-464d-aa9b-a811ee65e5d1" (UID: "21e86dc4-3382-464d-aa9b-a811ee65e5d1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.909303 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21e86dc4-3382-464d-aa9b-a811ee65e5d1" (UID: "21e86dc4-3382-464d-aa9b-a811ee65e5d1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.980185 4876 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.980238 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.980262 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mhn5\" (UniqueName: \"kubernetes.io/projected/21e86dc4-3382-464d-aa9b-a811ee65e5d1-kube-api-access-7mhn5\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.980273 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:11 crc kubenswrapper[4876]: I1205 08:57:11.980285 4876 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/21e86dc4-3382-464d-aa9b-a811ee65e5d1-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.278470 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" event={"ID":"21e86dc4-3382-464d-aa9b-a811ee65e5d1","Type":"ContainerDied","Data":"8867f09788afd50df8feeabe181267d592fbe32296df7f1c6acb6ba4e2183687"} Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.278512 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8867f09788afd50df8feeabe181267d592fbe32296df7f1c6acb6ba4e2183687" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.278567 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.377438 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz"] Dec 05 08:57:12 crc kubenswrapper[4876]: E1205 08:57:12.377853 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e86dc4-3382-464d-aa9b-a811ee65e5d1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.377874 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e86dc4-3382-464d-aa9b-a811ee65e5d1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.378105 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e86dc4-3382-464d-aa9b-a811ee65e5d1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.378792 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.380574 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.380824 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.380860 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.380998 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.381046 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.381282 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.381341 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.386710 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.386752 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj9dt\" (UniqueName: \"kubernetes.io/projected/d5afc65a-8adf-4aa6-a348-469c81b51bad-kube-api-access-wj9dt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.386781 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.386819 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.386838 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.387206 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.387271 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.387293 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.387509 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.401872 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz"] Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489434 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489488 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489516 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj9dt\" (UniqueName: \"kubernetes.io/projected/d5afc65a-8adf-4aa6-a348-469c81b51bad-kube-api-access-wj9dt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489550 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489597 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489614 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489645 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489706 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.489726 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.490892 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.495165 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.495328 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.495547 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.495635 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.495749 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.497603 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.498819 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.511821 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj9dt\" (UniqueName: \"kubernetes.io/projected/d5afc65a-8adf-4aa6-a348-469c81b51bad-kube-api-access-wj9dt\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5hxqz\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:12 crc kubenswrapper[4876]: I1205 08:57:12.697706 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:57:13 crc kubenswrapper[4876]: I1205 08:57:13.252311 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz"] Dec 05 08:57:13 crc kubenswrapper[4876]: I1205 08:57:13.263435 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 08:57:13 crc kubenswrapper[4876]: I1205 08:57:13.286968 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" event={"ID":"d5afc65a-8adf-4aa6-a348-469c81b51bad","Type":"ContainerStarted","Data":"6d13c5fca2683d67e018d006f61282baff8c466149ee7fe495c79a0f70249a65"} Dec 05 08:57:14 crc kubenswrapper[4876]: I1205 08:57:14.295740 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" event={"ID":"d5afc65a-8adf-4aa6-a348-469c81b51bad","Type":"ContainerStarted","Data":"1b77264bcfd34050111869dcfb1c7661ab92203e82d4696d2c240e821783a830"} Dec 05 08:57:14 crc kubenswrapper[4876]: I1205 08:57:14.320325 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" podStartSLOduration=1.866757923 podStartE2EDuration="2.320305642s" podCreationTimestamp="2025-12-05 08:57:12 +0000 UTC" firstStartedPulling="2025-12-05 08:57:13.263199721 +0000 UTC m=+2497.751864343" lastFinishedPulling="2025-12-05 08:57:13.71674744 +0000 UTC m=+2498.205412062" observedRunningTime="2025-12-05 08:57:14.314482123 +0000 UTC m=+2498.803146765" watchObservedRunningTime="2025-12-05 08:57:14.320305642 +0000 UTC m=+2498.808970264" Dec 05 08:57:19 crc kubenswrapper[4876]: I1205 08:57:19.823910 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:57:19 crc kubenswrapper[4876]: E1205 08:57:19.824812 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:57:34 crc kubenswrapper[4876]: I1205 08:57:34.824299 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:57:34 crc kubenswrapper[4876]: E1205 08:57:34.825055 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 08:57:47 crc kubenswrapper[4876]: I1205 08:57:47.825710 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 08:57:48 crc kubenswrapper[4876]: I1205 08:57:48.610258 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"c9e02980657b56eeb9e3238767783325bfa298fc32f0e06f3f0b7935ba1bf517"} Dec 05 08:59:48 crc kubenswrapper[4876]: I1205 08:59:48.675857 4876 generic.go:334] "Generic (PLEG): container finished" podID="d5afc65a-8adf-4aa6-a348-469c81b51bad" containerID="1b77264bcfd34050111869dcfb1c7661ab92203e82d4696d2c240e821783a830" exitCode=0 Dec 05 08:59:48 crc kubenswrapper[4876]: I1205 08:59:48.676013 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" event={"ID":"d5afc65a-8adf-4aa6-a348-469c81b51bad","Type":"ContainerDied","Data":"1b77264bcfd34050111869dcfb1c7661ab92203e82d4696d2c240e821783a830"} Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.177046 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277136 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-ssh-key\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277381 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-1\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277434 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-inventory\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277468 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-0\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277533 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-combined-ca-bundle\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277552 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-1\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277605 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-extra-config-0\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277626 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj9dt\" (UniqueName: \"kubernetes.io/projected/d5afc65a-8adf-4aa6-a348-469c81b51bad-kube-api-access-wj9dt\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.277650 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-0\") pod \"d5afc65a-8adf-4aa6-a348-469c81b51bad\" (UID: \"d5afc65a-8adf-4aa6-a348-469c81b51bad\") " Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.283177 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5afc65a-8adf-4aa6-a348-469c81b51bad-kube-api-access-wj9dt" (OuterVolumeSpecName: "kube-api-access-wj9dt") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "kube-api-access-wj9dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.301974 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.307422 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-inventory" (OuterVolumeSpecName: "inventory") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.311201 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.314393 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.314448 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.314663 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.316882 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.319608 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "d5afc65a-8adf-4aa6-a348-469c81b51bad" (UID: "d5afc65a-8adf-4aa6-a348-469c81b51bad"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380167 4876 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380198 4876 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380207 4876 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380218 4876 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380227 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj9dt\" (UniqueName: \"kubernetes.io/projected/d5afc65a-8adf-4aa6-a348-469c81b51bad-kube-api-access-wj9dt\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380236 4876 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380246 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380255 4876 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.380263 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5afc65a-8adf-4aa6-a348-469c81b51bad-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.697027 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" event={"ID":"d5afc65a-8adf-4aa6-a348-469c81b51bad","Type":"ContainerDied","Data":"6d13c5fca2683d67e018d006f61282baff8c466149ee7fe495c79a0f70249a65"} Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.697081 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d13c5fca2683d67e018d006f61282baff8c466149ee7fe495c79a0f70249a65" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.697112 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5hxqz" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.820400 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n"] Dec 05 08:59:50 crc kubenswrapper[4876]: E1205 08:59:50.821013 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5afc65a-8adf-4aa6-a348-469c81b51bad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.821040 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5afc65a-8adf-4aa6-a348-469c81b51bad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.821506 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5afc65a-8adf-4aa6-a348-469c81b51bad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.822498 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.824529 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.826572 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.827095 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.827288 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6tszg" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.827629 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.857096 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n"] Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.988813 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.988933 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.988957 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8rgk\" (UniqueName: \"kubernetes.io/projected/611e9ca4-b50e-415e-bf76-ec70f386af72-kube-api-access-c8rgk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.988979 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.989003 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.989234 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:50 crc kubenswrapper[4876]: I1205 08:59:50.989319 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.090970 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.091015 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.091066 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.091133 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.091157 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8rgk\" (UniqueName: \"kubernetes.io/projected/611e9ca4-b50e-415e-bf76-ec70f386af72-kube-api-access-c8rgk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.091181 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.091210 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.096397 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.096622 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.096766 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.096864 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.097248 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.108808 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.109129 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8rgk\" (UniqueName: \"kubernetes.io/projected/611e9ca4-b50e-415e-bf76-ec70f386af72-kube-api-access-c8rgk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-zc47n\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.156568 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.677868 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n"] Dec 05 08:59:51 crc kubenswrapper[4876]: I1205 08:59:51.705583 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" event={"ID":"611e9ca4-b50e-415e-bf76-ec70f386af72","Type":"ContainerStarted","Data":"fdcab26953a39a7a73c79ce9f78f05f0c86a2282d238be66d88226999a42f775"} Dec 05 08:59:52 crc kubenswrapper[4876]: I1205 08:59:52.715976 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" event={"ID":"611e9ca4-b50e-415e-bf76-ec70f386af72","Type":"ContainerStarted","Data":"babbb89e01b50c11be123f7e6ccda01dbd68d8cc7da408a0f1b9fc1b67e24615"} Dec 05 08:59:52 crc kubenswrapper[4876]: I1205 08:59:52.740160 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" podStartSLOduration=2.207775385 podStartE2EDuration="2.740136494s" podCreationTimestamp="2025-12-05 08:59:50 +0000 UTC" firstStartedPulling="2025-12-05 08:59:51.686731568 +0000 UTC m=+2656.175396180" lastFinishedPulling="2025-12-05 08:59:52.219092667 +0000 UTC m=+2656.707757289" observedRunningTime="2025-12-05 08:59:52.733919843 +0000 UTC m=+2657.222584465" watchObservedRunningTime="2025-12-05 08:59:52.740136494 +0000 UTC m=+2657.228801126" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.337983 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mlqrc"] Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.340866 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.356754 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mlqrc"] Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.436538 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-catalog-content\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.436677 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq89l\" (UniqueName: \"kubernetes.io/projected/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-kube-api-access-lq89l\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.436708 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-utilities\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.538498 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq89l\" (UniqueName: \"kubernetes.io/projected/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-kube-api-access-lq89l\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.540423 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-utilities\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.541474 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-catalog-content\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.541141 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-utilities\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.541836 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-catalog-content\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.562825 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq89l\" (UniqueName: \"kubernetes.io/projected/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-kube-api-access-lq89l\") pod \"community-operators-mlqrc\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:53 crc kubenswrapper[4876]: I1205 08:59:53.678975 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlqrc" Dec 05 08:59:54 crc kubenswrapper[4876]: I1205 08:59:54.213407 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mlqrc"] Dec 05 08:59:54 crc kubenswrapper[4876]: I1205 08:59:54.734053 4876 generic.go:334] "Generic (PLEG): container finished" podID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerID="3ae3ac28729c6b716f23b951a62317765e5e9a8db2bc23aac405ffa1b40e142a" exitCode=0 Dec 05 08:59:54 crc kubenswrapper[4876]: I1205 08:59:54.734171 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlqrc" event={"ID":"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29","Type":"ContainerDied","Data":"3ae3ac28729c6b716f23b951a62317765e5e9a8db2bc23aac405ffa1b40e142a"} Dec 05 08:59:54 crc kubenswrapper[4876]: I1205 08:59:54.734241 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlqrc" event={"ID":"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29","Type":"ContainerStarted","Data":"3fa888a8c97d925ad93da7c9dab58a5d13be830f573b5dcd6007193473ae63c3"} Dec 05 08:59:56 crc kubenswrapper[4876]: I1205 08:59:56.760168 4876 generic.go:334] "Generic (PLEG): container finished" podID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerID="2417e1525eff008562f3d3c7459caf29e986f337342ce4b523b6d152a1a09362" exitCode=0 Dec 05 08:59:56 crc kubenswrapper[4876]: I1205 08:59:56.760358 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlqrc" event={"ID":"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29","Type":"ContainerDied","Data":"2417e1525eff008562f3d3c7459caf29e986f337342ce4b523b6d152a1a09362"} Dec 05 08:59:57 crc kubenswrapper[4876]: I1205 08:59:57.771125 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlqrc" event={"ID":"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29","Type":"ContainerStarted","Data":"eeaa5db36b46e0c537d14a3ab30bb615b3127118cf8079863f5db65c301e66e7"} Dec 05 08:59:57 crc kubenswrapper[4876]: I1205 08:59:57.797584 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mlqrc" podStartSLOduration=2.398221544 podStartE2EDuration="4.797561452s" podCreationTimestamp="2025-12-05 08:59:53 +0000 UTC" firstStartedPulling="2025-12-05 08:59:54.737234219 +0000 UTC m=+2659.225898861" lastFinishedPulling="2025-12-05 08:59:57.136574127 +0000 UTC m=+2661.625238769" observedRunningTime="2025-12-05 08:59:57.793526331 +0000 UTC m=+2662.282190963" watchObservedRunningTime="2025-12-05 08:59:57.797561452 +0000 UTC m=+2662.286226074" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.134869 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk"] Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.138087 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.143255 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.143580 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.145451 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk"] Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.266778 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-config-volume\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.266840 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-secret-volume\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.266895 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jr9m\" (UniqueName: \"kubernetes.io/projected/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-kube-api-access-4jr9m\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.368811 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-config-volume\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.368866 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-secret-volume\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.368922 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jr9m\" (UniqueName: \"kubernetes.io/projected/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-kube-api-access-4jr9m\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.369880 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-config-volume\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.375392 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-secret-volume\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.386163 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jr9m\" (UniqueName: \"kubernetes.io/projected/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-kube-api-access-4jr9m\") pod \"collect-profiles-29415420-s7tqk\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.469671 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:00 crc kubenswrapper[4876]: I1205 09:00:00.913926 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk"] Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.351051 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bdm5m"] Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.353636 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.364442 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdm5m"] Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.389588 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-utilities\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.389641 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb4gk\" (UniqueName: \"kubernetes.io/projected/024ef427-ea1b-499e-ac06-92a742d38d4e-kube-api-access-gb4gk\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.389692 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-catalog-content\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.491130 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-utilities\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.491201 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb4gk\" (UniqueName: \"kubernetes.io/projected/024ef427-ea1b-499e-ac06-92a742d38d4e-kube-api-access-gb4gk\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.491281 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-catalog-content\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.491871 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-catalog-content\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.492211 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-utilities\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.516637 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb4gk\" (UniqueName: \"kubernetes.io/projected/024ef427-ea1b-499e-ac06-92a742d38d4e-kube-api-access-gb4gk\") pod \"redhat-operators-bdm5m\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.721490 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.806703 4876 generic.go:334] "Generic (PLEG): container finished" podID="514fa719-bc4c-4f5d-b78d-c0e0dbd43835" containerID="5cf90f253bf5621898699f9d39b2535dd9e7eb89eaea06eb2af37415c6288c1b" exitCode=0 Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.818726 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" event={"ID":"514fa719-bc4c-4f5d-b78d-c0e0dbd43835","Type":"ContainerDied","Data":"5cf90f253bf5621898699f9d39b2535dd9e7eb89eaea06eb2af37415c6288c1b"} Dec 05 09:00:01 crc kubenswrapper[4876]: I1205 09:00:01.818787 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" event={"ID":"514fa719-bc4c-4f5d-b78d-c0e0dbd43835","Type":"ContainerStarted","Data":"6cd7ab0c8add70fb3d2e6a24488dcf9797810476bf2ad6dc8dee34fe8713f33f"} Dec 05 09:00:02 crc kubenswrapper[4876]: I1205 09:00:02.222798 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdm5m"] Dec 05 09:00:02 crc kubenswrapper[4876]: W1205 09:00:02.226113 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod024ef427_ea1b_499e_ac06_92a742d38d4e.slice/crio-a2429da1f5d5f643f98eac8e2a30f20e9fd8cd148f683714b0d72003f1f5fd0b WatchSource:0}: Error finding container a2429da1f5d5f643f98eac8e2a30f20e9fd8cd148f683714b0d72003f1f5fd0b: Status 404 returned error can't find the container with id a2429da1f5d5f643f98eac8e2a30f20e9fd8cd148f683714b0d72003f1f5fd0b Dec 05 09:00:02 crc kubenswrapper[4876]: I1205 09:00:02.816464 4876 generic.go:334] "Generic (PLEG): container finished" podID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerID="b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65" exitCode=0 Dec 05 09:00:02 crc kubenswrapper[4876]: I1205 09:00:02.816543 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdm5m" event={"ID":"024ef427-ea1b-499e-ac06-92a742d38d4e","Type":"ContainerDied","Data":"b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65"} Dec 05 09:00:02 crc kubenswrapper[4876]: I1205 09:00:02.816591 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdm5m" event={"ID":"024ef427-ea1b-499e-ac06-92a742d38d4e","Type":"ContainerStarted","Data":"a2429da1f5d5f643f98eac8e2a30f20e9fd8cd148f683714b0d72003f1f5fd0b"} Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.193645 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.325469 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jr9m\" (UniqueName: \"kubernetes.io/projected/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-kube-api-access-4jr9m\") pod \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.325651 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-config-volume\") pod \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.325739 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-secret-volume\") pod \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\" (UID: \"514fa719-bc4c-4f5d-b78d-c0e0dbd43835\") " Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.327519 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-config-volume" (OuterVolumeSpecName: "config-volume") pod "514fa719-bc4c-4f5d-b78d-c0e0dbd43835" (UID: "514fa719-bc4c-4f5d-b78d-c0e0dbd43835"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.331635 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-kube-api-access-4jr9m" (OuterVolumeSpecName: "kube-api-access-4jr9m") pod "514fa719-bc4c-4f5d-b78d-c0e0dbd43835" (UID: "514fa719-bc4c-4f5d-b78d-c0e0dbd43835"). InnerVolumeSpecName "kube-api-access-4jr9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.338893 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "514fa719-bc4c-4f5d-b78d-c0e0dbd43835" (UID: "514fa719-bc4c-4f5d-b78d-c0e0dbd43835"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.428510 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jr9m\" (UniqueName: \"kubernetes.io/projected/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-kube-api-access-4jr9m\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.428560 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.428574 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/514fa719-bc4c-4f5d-b78d-c0e0dbd43835-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.679665 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mlqrc" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.679942 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mlqrc" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.726734 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mlqrc" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.828535 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.836346 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-s7tqk" event={"ID":"514fa719-bc4c-4f5d-b78d-c0e0dbd43835","Type":"ContainerDied","Data":"6cd7ab0c8add70fb3d2e6a24488dcf9797810476bf2ad6dc8dee34fe8713f33f"} Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.836385 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cd7ab0c8add70fb3d2e6a24488dcf9797810476bf2ad6dc8dee34fe8713f33f" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.902397 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mlqrc" Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.918112 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt"] Dec 05 09:00:03 crc kubenswrapper[4876]: I1205 09:00:03.929962 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415375-796gt"] Dec 05 09:00:04 crc kubenswrapper[4876]: I1205 09:00:04.837482 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdm5m" event={"ID":"024ef427-ea1b-499e-ac06-92a742d38d4e","Type":"ContainerStarted","Data":"1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52"} Dec 05 09:00:05 crc kubenswrapper[4876]: I1205 09:00:05.925833 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5742635-f4b3-47ee-88be-d59720792b25" path="/var/lib/kubelet/pods/b5742635-f4b3-47ee-88be-d59720792b25/volumes" Dec 05 09:00:06 crc kubenswrapper[4876]: I1205 09:00:06.131081 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mlqrc"] Dec 05 09:00:06 crc kubenswrapper[4876]: I1205 09:00:06.131352 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mlqrc" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerName="registry-server" containerID="cri-o://eeaa5db36b46e0c537d14a3ab30bb615b3127118cf8079863f5db65c301e66e7" gracePeriod=2 Dec 05 09:00:06 crc kubenswrapper[4876]: I1205 09:00:06.860707 4876 generic.go:334] "Generic (PLEG): container finished" podID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerID="1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52" exitCode=0 Dec 05 09:00:06 crc kubenswrapper[4876]: I1205 09:00:06.860753 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdm5m" event={"ID":"024ef427-ea1b-499e-ac06-92a742d38d4e","Type":"ContainerDied","Data":"1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52"} Dec 05 09:00:07 crc kubenswrapper[4876]: I1205 09:00:07.891217 4876 generic.go:334] "Generic (PLEG): container finished" podID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerID="eeaa5db36b46e0c537d14a3ab30bb615b3127118cf8079863f5db65c301e66e7" exitCode=0 Dec 05 09:00:07 crc kubenswrapper[4876]: I1205 09:00:07.891618 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlqrc" event={"ID":"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29","Type":"ContainerDied","Data":"eeaa5db36b46e0c537d14a3ab30bb615b3127118cf8079863f5db65c301e66e7"} Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.213621 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.213883 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.392798 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlqrc" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.523194 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq89l\" (UniqueName: \"kubernetes.io/projected/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-kube-api-access-lq89l\") pod \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.523282 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-utilities\") pod \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.523411 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-catalog-content\") pod \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\" (UID: \"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29\") " Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.525321 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-utilities" (OuterVolumeSpecName: "utilities") pod "df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" (UID: "df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.530643 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-kube-api-access-lq89l" (OuterVolumeSpecName: "kube-api-access-lq89l") pod "df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" (UID: "df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29"). InnerVolumeSpecName "kube-api-access-lq89l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.578233 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" (UID: "df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.625703 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.626017 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.626116 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq89l\" (UniqueName: \"kubernetes.io/projected/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29-kube-api-access-lq89l\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.901020 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdm5m" event={"ID":"024ef427-ea1b-499e-ac06-92a742d38d4e","Type":"ContainerStarted","Data":"408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41"} Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.904728 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mlqrc" event={"ID":"df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29","Type":"ContainerDied","Data":"3fa888a8c97d925ad93da7c9dab58a5d13be830f573b5dcd6007193473ae63c3"} Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.904768 4876 scope.go:117] "RemoveContainer" containerID="eeaa5db36b46e0c537d14a3ab30bb615b3127118cf8079863f5db65c301e66e7" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.904782 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mlqrc" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.935924 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bdm5m" podStartSLOduration=3.019254134 podStartE2EDuration="7.935886031s" podCreationTimestamp="2025-12-05 09:00:01 +0000 UTC" firstStartedPulling="2025-12-05 09:00:02.818054004 +0000 UTC m=+2667.306718636" lastFinishedPulling="2025-12-05 09:00:07.734685901 +0000 UTC m=+2672.223350533" observedRunningTime="2025-12-05 09:00:08.922613956 +0000 UTC m=+2673.411278578" watchObservedRunningTime="2025-12-05 09:00:08.935886031 +0000 UTC m=+2673.424550653" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.937265 4876 scope.go:117] "RemoveContainer" containerID="2417e1525eff008562f3d3c7459caf29e986f337342ce4b523b6d152a1a09362" Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.945378 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mlqrc"] Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.955808 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mlqrc"] Dec 05 09:00:08 crc kubenswrapper[4876]: I1205 09:00:08.977412 4876 scope.go:117] "RemoveContainer" containerID="3ae3ac28729c6b716f23b951a62317765e5e9a8db2bc23aac405ffa1b40e142a" Dec 05 09:00:09 crc kubenswrapper[4876]: I1205 09:00:09.836823 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" path="/var/lib/kubelet/pods/df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29/volumes" Dec 05 09:00:11 crc kubenswrapper[4876]: I1205 09:00:11.722389 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:11 crc kubenswrapper[4876]: I1205 09:00:11.722794 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:12 crc kubenswrapper[4876]: I1205 09:00:12.767684 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bdm5m" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="registry-server" probeResult="failure" output=< Dec 05 09:00:12 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 09:00:12 crc kubenswrapper[4876]: > Dec 05 09:00:21 crc kubenswrapper[4876]: I1205 09:00:21.776585 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:21 crc kubenswrapper[4876]: I1205 09:00:21.842107 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:22 crc kubenswrapper[4876]: I1205 09:00:22.036028 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdm5m"] Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.050283 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bdm5m" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="registry-server" containerID="cri-o://408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41" gracePeriod=2 Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.533174 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.600518 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb4gk\" (UniqueName: \"kubernetes.io/projected/024ef427-ea1b-499e-ac06-92a742d38d4e-kube-api-access-gb4gk\") pod \"024ef427-ea1b-499e-ac06-92a742d38d4e\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.600672 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-utilities\") pod \"024ef427-ea1b-499e-ac06-92a742d38d4e\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.600700 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-catalog-content\") pod \"024ef427-ea1b-499e-ac06-92a742d38d4e\" (UID: \"024ef427-ea1b-499e-ac06-92a742d38d4e\") " Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.601536 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-utilities" (OuterVolumeSpecName: "utilities") pod "024ef427-ea1b-499e-ac06-92a742d38d4e" (UID: "024ef427-ea1b-499e-ac06-92a742d38d4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.606132 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/024ef427-ea1b-499e-ac06-92a742d38d4e-kube-api-access-gb4gk" (OuterVolumeSpecName: "kube-api-access-gb4gk") pod "024ef427-ea1b-499e-ac06-92a742d38d4e" (UID: "024ef427-ea1b-499e-ac06-92a742d38d4e"). InnerVolumeSpecName "kube-api-access-gb4gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.703211 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb4gk\" (UniqueName: \"kubernetes.io/projected/024ef427-ea1b-499e-ac06-92a742d38d4e-kube-api-access-gb4gk\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.703248 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.726583 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "024ef427-ea1b-499e-ac06-92a742d38d4e" (UID: "024ef427-ea1b-499e-ac06-92a742d38d4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:00:23 crc kubenswrapper[4876]: I1205 09:00:23.804835 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/024ef427-ea1b-499e-ac06-92a742d38d4e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.061624 4876 generic.go:334] "Generic (PLEG): container finished" podID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerID="408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41" exitCode=0 Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.061670 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdm5m" event={"ID":"024ef427-ea1b-499e-ac06-92a742d38d4e","Type":"ContainerDied","Data":"408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41"} Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.061696 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdm5m" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.061709 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdm5m" event={"ID":"024ef427-ea1b-499e-ac06-92a742d38d4e","Type":"ContainerDied","Data":"a2429da1f5d5f643f98eac8e2a30f20e9fd8cd148f683714b0d72003f1f5fd0b"} Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.061730 4876 scope.go:117] "RemoveContainer" containerID="408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.095580 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdm5m"] Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.101042 4876 scope.go:117] "RemoveContainer" containerID="1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.105329 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bdm5m"] Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.126580 4876 scope.go:117] "RemoveContainer" containerID="b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.162749 4876 scope.go:117] "RemoveContainer" containerID="408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41" Dec 05 09:00:24 crc kubenswrapper[4876]: E1205 09:00:24.163260 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41\": container with ID starting with 408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41 not found: ID does not exist" containerID="408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.163333 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41"} err="failed to get container status \"408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41\": rpc error: code = NotFound desc = could not find container \"408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41\": container with ID starting with 408c332878ce802523abccd0d0e40f3744b51b60de442cb3e26f12050e4fee41 not found: ID does not exist" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.163402 4876 scope.go:117] "RemoveContainer" containerID="1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52" Dec 05 09:00:24 crc kubenswrapper[4876]: E1205 09:00:24.163998 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52\": container with ID starting with 1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52 not found: ID does not exist" containerID="1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.164079 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52"} err="failed to get container status \"1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52\": rpc error: code = NotFound desc = could not find container \"1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52\": container with ID starting with 1f10d5249ecfc4ad5ecfd27da57e60cd20ef659c378f8526844459528de75e52 not found: ID does not exist" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.164116 4876 scope.go:117] "RemoveContainer" containerID="b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65" Dec 05 09:00:24 crc kubenswrapper[4876]: E1205 09:00:24.164448 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65\": container with ID starting with b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65 not found: ID does not exist" containerID="b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65" Dec 05 09:00:24 crc kubenswrapper[4876]: I1205 09:00:24.164485 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65"} err="failed to get container status \"b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65\": rpc error: code = NotFound desc = could not find container \"b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65\": container with ID starting with b30f4cde8ff540773201fcc4e28cc63fbb1830ae485172cf885bb43536132e65 not found: ID does not exist" Dec 05 09:00:25 crc kubenswrapper[4876]: I1205 09:00:25.839119 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" path="/var/lib/kubelet/pods/024ef427-ea1b-499e-ac06-92a742d38d4e/volumes" Dec 05 09:00:38 crc kubenswrapper[4876]: I1205 09:00:38.214424 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:00:38 crc kubenswrapper[4876]: I1205 09:00:38.214949 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:00:42 crc kubenswrapper[4876]: I1205 09:00:42.064730 4876 scope.go:117] "RemoveContainer" containerID="efb2cedad915817582bbb028575ecd4ad526db2eb7c49cf2eb674d1c19766ef6" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.637450 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2llmg"] Dec 05 09:00:48 crc kubenswrapper[4876]: E1205 09:00:48.638494 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="extract-utilities" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638512 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="extract-utilities" Dec 05 09:00:48 crc kubenswrapper[4876]: E1205 09:00:48.638533 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerName="extract-utilities" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638541 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerName="extract-utilities" Dec 05 09:00:48 crc kubenswrapper[4876]: E1205 09:00:48.638556 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="registry-server" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638564 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="registry-server" Dec 05 09:00:48 crc kubenswrapper[4876]: E1205 09:00:48.638577 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerName="registry-server" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638585 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerName="registry-server" Dec 05 09:00:48 crc kubenswrapper[4876]: E1205 09:00:48.638605 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="extract-content" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638613 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="extract-content" Dec 05 09:00:48 crc kubenswrapper[4876]: E1205 09:00:48.638646 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514fa719-bc4c-4f5d-b78d-c0e0dbd43835" containerName="collect-profiles" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638669 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="514fa719-bc4c-4f5d-b78d-c0e0dbd43835" containerName="collect-profiles" Dec 05 09:00:48 crc kubenswrapper[4876]: E1205 09:00:48.638685 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerName="extract-content" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638692 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerName="extract-content" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638924 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="024ef427-ea1b-499e-ac06-92a742d38d4e" containerName="registry-server" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638941 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="514fa719-bc4c-4f5d-b78d-c0e0dbd43835" containerName="collect-profiles" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.638970 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="df0b8db0-d7c0-4e7f-b4a4-54b11fc29f29" containerName="registry-server" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.640476 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.650832 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llmg"] Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.697665 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-catalog-content\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.697738 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbbb5\" (UniqueName: \"kubernetes.io/projected/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-kube-api-access-rbbb5\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.697828 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-utilities\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.799578 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-catalog-content\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.799660 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbbb5\" (UniqueName: \"kubernetes.io/projected/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-kube-api-access-rbbb5\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.799711 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-utilities\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.800589 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-catalog-content\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.801130 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-utilities\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:48 crc kubenswrapper[4876]: I1205 09:00:48.819266 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbbb5\" (UniqueName: \"kubernetes.io/projected/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-kube-api-access-rbbb5\") pod \"redhat-marketplace-2llmg\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:49 crc kubenswrapper[4876]: I1205 09:00:49.005891 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:49 crc kubenswrapper[4876]: I1205 09:00:49.449512 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llmg"] Dec 05 09:00:50 crc kubenswrapper[4876]: I1205 09:00:50.292681 4876 generic.go:334] "Generic (PLEG): container finished" podID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerID="e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293" exitCode=0 Dec 05 09:00:50 crc kubenswrapper[4876]: I1205 09:00:50.292772 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llmg" event={"ID":"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d","Type":"ContainerDied","Data":"e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293"} Dec 05 09:00:50 crc kubenswrapper[4876]: I1205 09:00:50.293037 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llmg" event={"ID":"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d","Type":"ContainerStarted","Data":"57efd4679a0ed8391adfed9631305cb2f592a2da62c7b360bd6272f875cead94"} Dec 05 09:00:51 crc kubenswrapper[4876]: I1205 09:00:51.303285 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llmg" event={"ID":"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d","Type":"ContainerStarted","Data":"26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0"} Dec 05 09:00:52 crc kubenswrapper[4876]: I1205 09:00:52.313644 4876 generic.go:334] "Generic (PLEG): container finished" podID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerID="26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0" exitCode=0 Dec 05 09:00:52 crc kubenswrapper[4876]: I1205 09:00:52.313696 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llmg" event={"ID":"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d","Type":"ContainerDied","Data":"26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0"} Dec 05 09:00:53 crc kubenswrapper[4876]: I1205 09:00:53.324146 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llmg" event={"ID":"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d","Type":"ContainerStarted","Data":"b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7"} Dec 05 09:00:53 crc kubenswrapper[4876]: I1205 09:00:53.359785 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2llmg" podStartSLOduration=2.963604927 podStartE2EDuration="5.359765506s" podCreationTimestamp="2025-12-05 09:00:48 +0000 UTC" firstStartedPulling="2025-12-05 09:00:50.294435906 +0000 UTC m=+2714.783100528" lastFinishedPulling="2025-12-05 09:00:52.690596485 +0000 UTC m=+2717.179261107" observedRunningTime="2025-12-05 09:00:53.348566208 +0000 UTC m=+2717.837230830" watchObservedRunningTime="2025-12-05 09:00:53.359765506 +0000 UTC m=+2717.848430138" Dec 05 09:00:59 crc kubenswrapper[4876]: I1205 09:00:59.006441 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:59 crc kubenswrapper[4876]: I1205 09:00:59.006852 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:59 crc kubenswrapper[4876]: I1205 09:00:59.054689 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:59 crc kubenswrapper[4876]: I1205 09:00:59.427092 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:00:59 crc kubenswrapper[4876]: I1205 09:00:59.481130 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llmg"] Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.151428 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29415421-28x2l"] Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.153151 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.167018 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415421-28x2l"] Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.256174 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-combined-ca-bundle\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.256220 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-fernet-keys\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.256260 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-config-data\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.256500 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrwxb\" (UniqueName: \"kubernetes.io/projected/778ede62-2e7e-4967-a96a-a8073b375f18-kube-api-access-vrwxb\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.359443 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-combined-ca-bundle\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.359512 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-fernet-keys\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.359567 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-config-data\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.359643 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrwxb\" (UniqueName: \"kubernetes.io/projected/778ede62-2e7e-4967-a96a-a8073b375f18-kube-api-access-vrwxb\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.365486 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-combined-ca-bundle\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.365726 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-fernet-keys\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.366097 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-config-data\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.375579 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrwxb\" (UniqueName: \"kubernetes.io/projected/778ede62-2e7e-4967-a96a-a8073b375f18-kube-api-access-vrwxb\") pod \"keystone-cron-29415421-28x2l\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.476556 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:00 crc kubenswrapper[4876]: I1205 09:01:00.957081 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415421-28x2l"] Dec 05 09:01:01 crc kubenswrapper[4876]: I1205 09:01:01.396618 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415421-28x2l" event={"ID":"778ede62-2e7e-4967-a96a-a8073b375f18","Type":"ContainerStarted","Data":"65d3cbfddcf9d31cdd07d2cb1d9a961a69aa30db566f62a9191fe2ea9e4c160b"} Dec 05 09:01:01 crc kubenswrapper[4876]: I1205 09:01:01.396989 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415421-28x2l" event={"ID":"778ede62-2e7e-4967-a96a-a8073b375f18","Type":"ContainerStarted","Data":"69489f0564ddf81ec8c26d326f7c48d568f2d525202af6f52a6c75f84726437b"} Dec 05 09:01:01 crc kubenswrapper[4876]: I1205 09:01:01.396771 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2llmg" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerName="registry-server" containerID="cri-o://b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7" gracePeriod=2 Dec 05 09:01:01 crc kubenswrapper[4876]: I1205 09:01:01.443948 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29415421-28x2l" podStartSLOduration=1.443931975 podStartE2EDuration="1.443931975s" podCreationTimestamp="2025-12-05 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:01:01.441656661 +0000 UTC m=+2725.930321293" watchObservedRunningTime="2025-12-05 09:01:01.443931975 +0000 UTC m=+2725.932596607" Dec 05 09:01:01 crc kubenswrapper[4876]: I1205 09:01:01.890483 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.032349 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-catalog-content\") pod \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.032484 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-utilities\") pod \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.032513 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbbb5\" (UniqueName: \"kubernetes.io/projected/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-kube-api-access-rbbb5\") pod \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\" (UID: \"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d\") " Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.033894 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-utilities" (OuterVolumeSpecName: "utilities") pod "d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" (UID: "d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.037812 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-kube-api-access-rbbb5" (OuterVolumeSpecName: "kube-api-access-rbbb5") pod "d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" (UID: "d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d"). InnerVolumeSpecName "kube-api-access-rbbb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.039237 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.039262 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbbb5\" (UniqueName: \"kubernetes.io/projected/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-kube-api-access-rbbb5\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.052455 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" (UID: "d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.145270 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.408278 4876 generic.go:334] "Generic (PLEG): container finished" podID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerID="b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7" exitCode=0 Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.408384 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2llmg" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.408427 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llmg" event={"ID":"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d","Type":"ContainerDied","Data":"b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7"} Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.408456 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llmg" event={"ID":"d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d","Type":"ContainerDied","Data":"57efd4679a0ed8391adfed9631305cb2f592a2da62c7b360bd6272f875cead94"} Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.408472 4876 scope.go:117] "RemoveContainer" containerID="b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.437894 4876 scope.go:117] "RemoveContainer" containerID="26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.483194 4876 scope.go:117] "RemoveContainer" containerID="e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.483863 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llmg"] Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.492300 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llmg"] Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.522814 4876 scope.go:117] "RemoveContainer" containerID="b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7" Dec 05 09:01:02 crc kubenswrapper[4876]: E1205 09:01:02.523923 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7\": container with ID starting with b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7 not found: ID does not exist" containerID="b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.523971 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7"} err="failed to get container status \"b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7\": rpc error: code = NotFound desc = could not find container \"b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7\": container with ID starting with b8576cd1c19969540b0f342a54bfff7930a437d7cc1d31212aae932e997f4ca7 not found: ID does not exist" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.524008 4876 scope.go:117] "RemoveContainer" containerID="26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0" Dec 05 09:01:02 crc kubenswrapper[4876]: E1205 09:01:02.524571 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0\": container with ID starting with 26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0 not found: ID does not exist" containerID="26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.524604 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0"} err="failed to get container status \"26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0\": rpc error: code = NotFound desc = could not find container \"26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0\": container with ID starting with 26b90e7a4e05d125c6d3f3bf09f205d2455792e2f57b7dd64fa847c01ad606b0 not found: ID does not exist" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.524623 4876 scope.go:117] "RemoveContainer" containerID="e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293" Dec 05 09:01:02 crc kubenswrapper[4876]: E1205 09:01:02.525014 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293\": container with ID starting with e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293 not found: ID does not exist" containerID="e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293" Dec 05 09:01:02 crc kubenswrapper[4876]: I1205 09:01:02.525134 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293"} err="failed to get container status \"e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293\": rpc error: code = NotFound desc = could not find container \"e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293\": container with ID starting with e78ff215f10357450518e4cbf29ddc8e2da2cb5c9cef9d4867a8c8b6769d1293 not found: ID does not exist" Dec 05 09:01:03 crc kubenswrapper[4876]: I1205 09:01:03.420311 4876 generic.go:334] "Generic (PLEG): container finished" podID="778ede62-2e7e-4967-a96a-a8073b375f18" containerID="65d3cbfddcf9d31cdd07d2cb1d9a961a69aa30db566f62a9191fe2ea9e4c160b" exitCode=0 Dec 05 09:01:03 crc kubenswrapper[4876]: I1205 09:01:03.420408 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415421-28x2l" event={"ID":"778ede62-2e7e-4967-a96a-a8073b375f18","Type":"ContainerDied","Data":"65d3cbfddcf9d31cdd07d2cb1d9a961a69aa30db566f62a9191fe2ea9e4c160b"} Dec 05 09:01:03 crc kubenswrapper[4876]: I1205 09:01:03.835088 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" path="/var/lib/kubelet/pods/d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d/volumes" Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.764519 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.901170 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrwxb\" (UniqueName: \"kubernetes.io/projected/778ede62-2e7e-4967-a96a-a8073b375f18-kube-api-access-vrwxb\") pod \"778ede62-2e7e-4967-a96a-a8073b375f18\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.901794 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-combined-ca-bundle\") pod \"778ede62-2e7e-4967-a96a-a8073b375f18\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.902058 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-config-data\") pod \"778ede62-2e7e-4967-a96a-a8073b375f18\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.902356 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-fernet-keys\") pod \"778ede62-2e7e-4967-a96a-a8073b375f18\" (UID: \"778ede62-2e7e-4967-a96a-a8073b375f18\") " Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.906438 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "778ede62-2e7e-4967-a96a-a8073b375f18" (UID: "778ede62-2e7e-4967-a96a-a8073b375f18"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.906919 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778ede62-2e7e-4967-a96a-a8073b375f18-kube-api-access-vrwxb" (OuterVolumeSpecName: "kube-api-access-vrwxb") pod "778ede62-2e7e-4967-a96a-a8073b375f18" (UID: "778ede62-2e7e-4967-a96a-a8073b375f18"). InnerVolumeSpecName "kube-api-access-vrwxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.929076 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "778ede62-2e7e-4967-a96a-a8073b375f18" (UID: "778ede62-2e7e-4967-a96a-a8073b375f18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:04 crc kubenswrapper[4876]: I1205 09:01:04.949335 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-config-data" (OuterVolumeSpecName: "config-data") pod "778ede62-2e7e-4967-a96a-a8073b375f18" (UID: "778ede62-2e7e-4967-a96a-a8073b375f18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:01:05 crc kubenswrapper[4876]: I1205 09:01:05.006562 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrwxb\" (UniqueName: \"kubernetes.io/projected/778ede62-2e7e-4967-a96a-a8073b375f18-kube-api-access-vrwxb\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:05 crc kubenswrapper[4876]: I1205 09:01:05.006608 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:05 crc kubenswrapper[4876]: I1205 09:01:05.006625 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:05 crc kubenswrapper[4876]: I1205 09:01:05.006640 4876 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/778ede62-2e7e-4967-a96a-a8073b375f18-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 09:01:05 crc kubenswrapper[4876]: I1205 09:01:05.458396 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415421-28x2l" event={"ID":"778ede62-2e7e-4967-a96a-a8073b375f18","Type":"ContainerDied","Data":"69489f0564ddf81ec8c26d326f7c48d568f2d525202af6f52a6c75f84726437b"} Dec 05 09:01:05 crc kubenswrapper[4876]: I1205 09:01:05.458450 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69489f0564ddf81ec8c26d326f7c48d568f2d525202af6f52a6c75f84726437b" Dec 05 09:01:05 crc kubenswrapper[4876]: I1205 09:01:05.458479 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415421-28x2l" Dec 05 09:01:08 crc kubenswrapper[4876]: I1205 09:01:08.214441 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:01:08 crc kubenswrapper[4876]: I1205 09:01:08.214808 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:01:08 crc kubenswrapper[4876]: I1205 09:01:08.214868 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 09:01:08 crc kubenswrapper[4876]: I1205 09:01:08.215877 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9e02980657b56eeb9e3238767783325bfa298fc32f0e06f3f0b7935ba1bf517"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:01:08 crc kubenswrapper[4876]: I1205 09:01:08.215989 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://c9e02980657b56eeb9e3238767783325bfa298fc32f0e06f3f0b7935ba1bf517" gracePeriod=600 Dec 05 09:01:08 crc kubenswrapper[4876]: I1205 09:01:08.507913 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="c9e02980657b56eeb9e3238767783325bfa298fc32f0e06f3f0b7935ba1bf517" exitCode=0 Dec 05 09:01:08 crc kubenswrapper[4876]: I1205 09:01:08.507947 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"c9e02980657b56eeb9e3238767783325bfa298fc32f0e06f3f0b7935ba1bf517"} Dec 05 09:01:08 crc kubenswrapper[4876]: I1205 09:01:08.508328 4876 scope.go:117] "RemoveContainer" containerID="cd5678846d6aaf9e7d5f99d8de5ab9bc57fcd6b6796e7b6e915b75a3a5eac412" Dec 05 09:01:09 crc kubenswrapper[4876]: I1205 09:01:09.519307 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e"} Dec 05 09:02:01 crc kubenswrapper[4876]: I1205 09:02:01.661893 4876 generic.go:334] "Generic (PLEG): container finished" podID="611e9ca4-b50e-415e-bf76-ec70f386af72" containerID="babbb89e01b50c11be123f7e6ccda01dbd68d8cc7da408a0f1b9fc1b67e24615" exitCode=0 Dec 05 09:02:01 crc kubenswrapper[4876]: I1205 09:02:01.662024 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" event={"ID":"611e9ca4-b50e-415e-bf76-ec70f386af72","Type":"ContainerDied","Data":"babbb89e01b50c11be123f7e6ccda01dbd68d8cc7da408a0f1b9fc1b67e24615"} Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.117560 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.270344 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-0\") pod \"611e9ca4-b50e-415e-bf76-ec70f386af72\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.270395 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-telemetry-combined-ca-bundle\") pod \"611e9ca4-b50e-415e-bf76-ec70f386af72\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.270504 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ssh-key\") pod \"611e9ca4-b50e-415e-bf76-ec70f386af72\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.270617 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-1\") pod \"611e9ca4-b50e-415e-bf76-ec70f386af72\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.270659 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-2\") pod \"611e9ca4-b50e-415e-bf76-ec70f386af72\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.270693 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-inventory\") pod \"611e9ca4-b50e-415e-bf76-ec70f386af72\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.270741 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8rgk\" (UniqueName: \"kubernetes.io/projected/611e9ca4-b50e-415e-bf76-ec70f386af72-kube-api-access-c8rgk\") pod \"611e9ca4-b50e-415e-bf76-ec70f386af72\" (UID: \"611e9ca4-b50e-415e-bf76-ec70f386af72\") " Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.276070 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/611e9ca4-b50e-415e-bf76-ec70f386af72-kube-api-access-c8rgk" (OuterVolumeSpecName: "kube-api-access-c8rgk") pod "611e9ca4-b50e-415e-bf76-ec70f386af72" (UID: "611e9ca4-b50e-415e-bf76-ec70f386af72"). InnerVolumeSpecName "kube-api-access-c8rgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.285246 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "611e9ca4-b50e-415e-bf76-ec70f386af72" (UID: "611e9ca4-b50e-415e-bf76-ec70f386af72"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.301840 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "611e9ca4-b50e-415e-bf76-ec70f386af72" (UID: "611e9ca4-b50e-415e-bf76-ec70f386af72"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.303020 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "611e9ca4-b50e-415e-bf76-ec70f386af72" (UID: "611e9ca4-b50e-415e-bf76-ec70f386af72"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.303610 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "611e9ca4-b50e-415e-bf76-ec70f386af72" (UID: "611e9ca4-b50e-415e-bf76-ec70f386af72"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.308702 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-inventory" (OuterVolumeSpecName: "inventory") pod "611e9ca4-b50e-415e-bf76-ec70f386af72" (UID: "611e9ca4-b50e-415e-bf76-ec70f386af72"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.308827 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "611e9ca4-b50e-415e-bf76-ec70f386af72" (UID: "611e9ca4-b50e-415e-bf76-ec70f386af72"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.372498 4876 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.372531 4876 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.372544 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8rgk\" (UniqueName: \"kubernetes.io/projected/611e9ca4-b50e-415e-bf76-ec70f386af72-kube-api-access-c8rgk\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.372554 4876 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.372564 4876 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.372573 4876 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.372583 4876 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/611e9ca4-b50e-415e-bf76-ec70f386af72-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.686923 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" event={"ID":"611e9ca4-b50e-415e-bf76-ec70f386af72","Type":"ContainerDied","Data":"fdcab26953a39a7a73c79ce9f78f05f0c86a2282d238be66d88226999a42f775"} Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.687286 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdcab26953a39a7a73c79ce9f78f05f0c86a2282d238be66d88226999a42f775" Dec 05 09:02:03 crc kubenswrapper[4876]: I1205 09:02:03.686968 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-zc47n" Dec 05 09:03:08 crc kubenswrapper[4876]: I1205 09:03:08.213351 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:03:08 crc kubenswrapper[4876]: I1205 09:03:08.213752 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.234582 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g294b"] Dec 05 09:03:10 crc kubenswrapper[4876]: E1205 09:03:10.235254 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerName="extract-content" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.235275 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerName="extract-content" Dec 05 09:03:10 crc kubenswrapper[4876]: E1205 09:03:10.235296 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerName="registry-server" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.235306 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerName="registry-server" Dec 05 09:03:10 crc kubenswrapper[4876]: E1205 09:03:10.235325 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerName="extract-utilities" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.235332 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerName="extract-utilities" Dec 05 09:03:10 crc kubenswrapper[4876]: E1205 09:03:10.235342 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778ede62-2e7e-4967-a96a-a8073b375f18" containerName="keystone-cron" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.235349 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="778ede62-2e7e-4967-a96a-a8073b375f18" containerName="keystone-cron" Dec 05 09:03:10 crc kubenswrapper[4876]: E1205 09:03:10.235367 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611e9ca4-b50e-415e-bf76-ec70f386af72" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.235375 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="611e9ca4-b50e-415e-bf76-ec70f386af72" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.235557 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="778ede62-2e7e-4967-a96a-a8073b375f18" containerName="keystone-cron" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.235578 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="611e9ca4-b50e-415e-bf76-ec70f386af72" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.235595 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3ad06d3-0278-4b3b-a9d0-0f063bb2e38d" containerName="registry-server" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.236964 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.256274 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g294b"] Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.321584 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-utilities\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.321683 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-catalog-content\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.321750 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj5ch\" (UniqueName: \"kubernetes.io/projected/138b2d31-b807-4fba-837a-92b583581a0c-kube-api-access-wj5ch\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.423320 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj5ch\" (UniqueName: \"kubernetes.io/projected/138b2d31-b807-4fba-837a-92b583581a0c-kube-api-access-wj5ch\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.423481 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-utilities\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.423558 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-catalog-content\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.424089 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-utilities\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.424126 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-catalog-content\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.446347 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj5ch\" (UniqueName: \"kubernetes.io/projected/138b2d31-b807-4fba-837a-92b583581a0c-kube-api-access-wj5ch\") pod \"certified-operators-g294b\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:10 crc kubenswrapper[4876]: I1205 09:03:10.557105 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:11 crc kubenswrapper[4876]: I1205 09:03:11.059675 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g294b"] Dec 05 09:03:11 crc kubenswrapper[4876]: I1205 09:03:11.298000 4876 generic.go:334] "Generic (PLEG): container finished" podID="138b2d31-b807-4fba-837a-92b583581a0c" containerID="33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3" exitCode=0 Dec 05 09:03:11 crc kubenswrapper[4876]: I1205 09:03:11.298062 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g294b" event={"ID":"138b2d31-b807-4fba-837a-92b583581a0c","Type":"ContainerDied","Data":"33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3"} Dec 05 09:03:11 crc kubenswrapper[4876]: I1205 09:03:11.298242 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g294b" event={"ID":"138b2d31-b807-4fba-837a-92b583581a0c","Type":"ContainerStarted","Data":"f139dba42ca6032482f723aa77cabafd3b7952b8c195c11c098fc27330219669"} Dec 05 09:03:11 crc kubenswrapper[4876]: I1205 09:03:11.302165 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:03:12 crc kubenswrapper[4876]: I1205 09:03:12.308146 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g294b" event={"ID":"138b2d31-b807-4fba-837a-92b583581a0c","Type":"ContainerStarted","Data":"e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3"} Dec 05 09:03:13 crc kubenswrapper[4876]: I1205 09:03:13.317616 4876 generic.go:334] "Generic (PLEG): container finished" podID="138b2d31-b807-4fba-837a-92b583581a0c" containerID="e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3" exitCode=0 Dec 05 09:03:13 crc kubenswrapper[4876]: I1205 09:03:13.317665 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g294b" event={"ID":"138b2d31-b807-4fba-837a-92b583581a0c","Type":"ContainerDied","Data":"e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3"} Dec 05 09:03:14 crc kubenswrapper[4876]: I1205 09:03:14.327459 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g294b" event={"ID":"138b2d31-b807-4fba-837a-92b583581a0c","Type":"ContainerStarted","Data":"eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc"} Dec 05 09:03:14 crc kubenswrapper[4876]: I1205 09:03:14.347522 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g294b" podStartSLOduration=1.84445294 podStartE2EDuration="4.347502687s" podCreationTimestamp="2025-12-05 09:03:10 +0000 UTC" firstStartedPulling="2025-12-05 09:03:11.300182664 +0000 UTC m=+2855.788847286" lastFinishedPulling="2025-12-05 09:03:13.803232411 +0000 UTC m=+2858.291897033" observedRunningTime="2025-12-05 09:03:14.347249009 +0000 UTC m=+2858.835913651" watchObservedRunningTime="2025-12-05 09:03:14.347502687 +0000 UTC m=+2858.836167309" Dec 05 09:03:20 crc kubenswrapper[4876]: I1205 09:03:20.557960 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:20 crc kubenswrapper[4876]: I1205 09:03:20.558663 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:20 crc kubenswrapper[4876]: I1205 09:03:20.603894 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:21 crc kubenswrapper[4876]: I1205 09:03:21.442230 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:21 crc kubenswrapper[4876]: I1205 09:03:21.492358 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g294b"] Dec 05 09:03:23 crc kubenswrapper[4876]: I1205 09:03:23.408385 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g294b" podUID="138b2d31-b807-4fba-837a-92b583581a0c" containerName="registry-server" containerID="cri-o://eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc" gracePeriod=2 Dec 05 09:03:23 crc kubenswrapper[4876]: I1205 09:03:23.851843 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:23 crc kubenswrapper[4876]: I1205 09:03:23.977106 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-catalog-content\") pod \"138b2d31-b807-4fba-837a-92b583581a0c\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " Dec 05 09:03:23 crc kubenswrapper[4876]: I1205 09:03:23.977156 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-utilities\") pod \"138b2d31-b807-4fba-837a-92b583581a0c\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " Dec 05 09:03:23 crc kubenswrapper[4876]: I1205 09:03:23.977178 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj5ch\" (UniqueName: \"kubernetes.io/projected/138b2d31-b807-4fba-837a-92b583581a0c-kube-api-access-wj5ch\") pod \"138b2d31-b807-4fba-837a-92b583581a0c\" (UID: \"138b2d31-b807-4fba-837a-92b583581a0c\") " Dec 05 09:03:23 crc kubenswrapper[4876]: I1205 09:03:23.978219 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-utilities" (OuterVolumeSpecName: "utilities") pod "138b2d31-b807-4fba-837a-92b583581a0c" (UID: "138b2d31-b807-4fba-837a-92b583581a0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:03:23 crc kubenswrapper[4876]: I1205 09:03:23.979067 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:23 crc kubenswrapper[4876]: I1205 09:03:23.983062 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/138b2d31-b807-4fba-837a-92b583581a0c-kube-api-access-wj5ch" (OuterVolumeSpecName: "kube-api-access-wj5ch") pod "138b2d31-b807-4fba-837a-92b583581a0c" (UID: "138b2d31-b807-4fba-837a-92b583581a0c"). InnerVolumeSpecName "kube-api-access-wj5ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.027438 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "138b2d31-b807-4fba-837a-92b583581a0c" (UID: "138b2d31-b807-4fba-837a-92b583581a0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.080669 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/138b2d31-b807-4fba-837a-92b583581a0c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.080707 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj5ch\" (UniqueName: \"kubernetes.io/projected/138b2d31-b807-4fba-837a-92b583581a0c-kube-api-access-wj5ch\") on node \"crc\" DevicePath \"\"" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.420720 4876 generic.go:334] "Generic (PLEG): container finished" podID="138b2d31-b807-4fba-837a-92b583581a0c" containerID="eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc" exitCode=0 Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.420797 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g294b" event={"ID":"138b2d31-b807-4fba-837a-92b583581a0c","Type":"ContainerDied","Data":"eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc"} Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.420856 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g294b" event={"ID":"138b2d31-b807-4fba-837a-92b583581a0c","Type":"ContainerDied","Data":"f139dba42ca6032482f723aa77cabafd3b7952b8c195c11c098fc27330219669"} Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.420875 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g294b" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.420944 4876 scope.go:117] "RemoveContainer" containerID="eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.457625 4876 scope.go:117] "RemoveContainer" containerID="e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.465717 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g294b"] Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.480108 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g294b"] Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.483885 4876 scope.go:117] "RemoveContainer" containerID="33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.534660 4876 scope.go:117] "RemoveContainer" containerID="eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc" Dec 05 09:03:24 crc kubenswrapper[4876]: E1205 09:03:24.535112 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc\": container with ID starting with eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc not found: ID does not exist" containerID="eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.535151 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc"} err="failed to get container status \"eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc\": rpc error: code = NotFound desc = could not find container \"eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc\": container with ID starting with eb75440e2ce396e10b748619f30e7834aab57bf9b3c7445be2fc2cb7d6dcb5bc not found: ID does not exist" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.535170 4876 scope.go:117] "RemoveContainer" containerID="e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3" Dec 05 09:03:24 crc kubenswrapper[4876]: E1205 09:03:24.535714 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3\": container with ID starting with e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3 not found: ID does not exist" containerID="e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.535838 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3"} err="failed to get container status \"e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3\": rpc error: code = NotFound desc = could not find container \"e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3\": container with ID starting with e2deb65b45e6e10b5986c58d7901e07eae7c113293734b93f8fe4643495cd7f3 not found: ID does not exist" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.536011 4876 scope.go:117] "RemoveContainer" containerID="33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3" Dec 05 09:03:24 crc kubenswrapper[4876]: E1205 09:03:24.536416 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3\": container with ID starting with 33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3 not found: ID does not exist" containerID="33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3" Dec 05 09:03:24 crc kubenswrapper[4876]: I1205 09:03:24.536444 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3"} err="failed to get container status \"33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3\": rpc error: code = NotFound desc = could not find container \"33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3\": container with ID starting with 33622389d4e25ad42386c8879d32f7816972e5a2f5edf82fbea1ebf5fc433ca3 not found: ID does not exist" Dec 05 09:03:25 crc kubenswrapper[4876]: I1205 09:03:25.835234 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="138b2d31-b807-4fba-837a-92b583581a0c" path="/var/lib/kubelet/pods/138b2d31-b807-4fba-837a-92b583581a0c/volumes" Dec 05 09:03:38 crc kubenswrapper[4876]: I1205 09:03:38.213326 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:03:38 crc kubenswrapper[4876]: I1205 09:03:38.213837 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.213693 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.214228 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.214282 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.215203 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.215259 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" gracePeriod=600 Dec 05 09:04:08 crc kubenswrapper[4876]: E1205 09:04:08.343494 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.878233 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" exitCode=0 Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.878291 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e"} Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.878332 4876 scope.go:117] "RemoveContainer" containerID="c9e02980657b56eeb9e3238767783325bfa298fc32f0e06f3f0b7935ba1bf517" Dec 05 09:04:08 crc kubenswrapper[4876]: I1205 09:04:08.879202 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:04:08 crc kubenswrapper[4876]: E1205 09:04:08.879494 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:04:23 crc kubenswrapper[4876]: I1205 09:04:23.825098 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:04:23 crc kubenswrapper[4876]: E1205 09:04:23.826054 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:04:34 crc kubenswrapper[4876]: I1205 09:04:34.823765 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:04:34 crc kubenswrapper[4876]: E1205 09:04:34.824512 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:04:49 crc kubenswrapper[4876]: I1205 09:04:49.824456 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:04:49 crc kubenswrapper[4876]: E1205 09:04:49.825387 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:05:01 crc kubenswrapper[4876]: I1205 09:05:01.874847 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/manager/0.log" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.790455 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.791006 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="86ac21be-c7c2-4035-a32c-3f24715c3248" containerName="openstackclient" containerID="cri-o://48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46" gracePeriod=2 Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.799201 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.845202 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 09:05:03 crc kubenswrapper[4876]: E1205 09:05:03.845494 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138b2d31-b807-4fba-837a-92b583581a0c" containerName="extract-content" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.845510 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="138b2d31-b807-4fba-837a-92b583581a0c" containerName="extract-content" Dec 05 09:05:03 crc kubenswrapper[4876]: E1205 09:05:03.845538 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ac21be-c7c2-4035-a32c-3f24715c3248" containerName="openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.845545 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ac21be-c7c2-4035-a32c-3f24715c3248" containerName="openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: E1205 09:05:03.845573 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138b2d31-b807-4fba-837a-92b583581a0c" containerName="registry-server" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.845578 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="138b2d31-b807-4fba-837a-92b583581a0c" containerName="registry-server" Dec 05 09:05:03 crc kubenswrapper[4876]: E1205 09:05:03.845592 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="138b2d31-b807-4fba-837a-92b583581a0c" containerName="extract-utilities" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.845598 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="138b2d31-b807-4fba-837a-92b583581a0c" containerName="extract-utilities" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.845769 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="138b2d31-b807-4fba-837a-92b583581a0c" containerName="registry-server" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.845786 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ac21be-c7c2-4035-a32c-3f24715c3248" containerName="openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.846352 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.846475 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.863566 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk98c\" (UniqueName: \"kubernetes.io/projected/bf04a5a6-de08-4983-8986-c0c949090943-kube-api-access-fk98c\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.863718 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config-secret\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.863761 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.863800 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.965310 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config-secret\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.965738 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.965788 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.966656 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk98c\" (UniqueName: \"kubernetes.io/projected/bf04a5a6-de08-4983-8986-c0c949090943-kube-api-access-fk98c\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.967039 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.982430 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.982684 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config-secret\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:03 crc kubenswrapper[4876]: I1205 09:05:03.996487 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk98c\" (UniqueName: \"kubernetes.io/projected/bf04a5a6-de08-4983-8986-c0c949090943-kube-api-access-fk98c\") pod \"openstackclient\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " pod="openstack/openstackclient" Dec 05 09:05:04 crc kubenswrapper[4876]: I1205 09:05:04.165708 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:05:04 crc kubenswrapper[4876]: I1205 09:05:04.723928 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:05:04 crc kubenswrapper[4876]: W1205 09:05:04.726567 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf04a5a6_de08_4983_8986_c0c949090943.slice/crio-2c9f943dd978740796ae5db72442514cf48fccc4a7c71296d92fb1ee9dc4fe45 WatchSource:0}: Error finding container 2c9f943dd978740796ae5db72442514cf48fccc4a7c71296d92fb1ee9dc4fe45: Status 404 returned error can't find the container with id 2c9f943dd978740796ae5db72442514cf48fccc4a7c71296d92fb1ee9dc4fe45 Dec 05 09:05:04 crc kubenswrapper[4876]: I1205 09:05:04.824028 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:05:04 crc kubenswrapper[4876]: E1205 09:05:04.824621 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.139516 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-jljwh"] Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.141066 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.149372 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-6a58-account-create-update-txbfr"] Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.150683 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.152187 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.158492 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-6a58-account-create-update-txbfr"] Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.200650 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jljwh"] Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.291678 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0f71887-ac32-4542-9a91-d1e350331dbc-operator-scripts\") pod \"aodh-db-create-jljwh\" (UID: \"e0f71887-ac32-4542-9a91-d1e350331dbc\") " pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.291734 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316c29a1-4346-4be0-be25-aa29a1a9dd0d-operator-scripts\") pod \"aodh-6a58-account-create-update-txbfr\" (UID: \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\") " pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.291948 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdqgb\" (UniqueName: \"kubernetes.io/projected/316c29a1-4346-4be0-be25-aa29a1a9dd0d-kube-api-access-jdqgb\") pod \"aodh-6a58-account-create-update-txbfr\" (UID: \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\") " pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.292018 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7558z\" (UniqueName: \"kubernetes.io/projected/e0f71887-ac32-4542-9a91-d1e350331dbc-kube-api-access-7558z\") pod \"aodh-db-create-jljwh\" (UID: \"e0f71887-ac32-4542-9a91-d1e350331dbc\") " pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.394135 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7558z\" (UniqueName: \"kubernetes.io/projected/e0f71887-ac32-4542-9a91-d1e350331dbc-kube-api-access-7558z\") pod \"aodh-db-create-jljwh\" (UID: \"e0f71887-ac32-4542-9a91-d1e350331dbc\") " pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.394269 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0f71887-ac32-4542-9a91-d1e350331dbc-operator-scripts\") pod \"aodh-db-create-jljwh\" (UID: \"e0f71887-ac32-4542-9a91-d1e350331dbc\") " pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.394296 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316c29a1-4346-4be0-be25-aa29a1a9dd0d-operator-scripts\") pod \"aodh-6a58-account-create-update-txbfr\" (UID: \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\") " pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.394446 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdqgb\" (UniqueName: \"kubernetes.io/projected/316c29a1-4346-4be0-be25-aa29a1a9dd0d-kube-api-access-jdqgb\") pod \"aodh-6a58-account-create-update-txbfr\" (UID: \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\") " pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.395234 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0f71887-ac32-4542-9a91-d1e350331dbc-operator-scripts\") pod \"aodh-db-create-jljwh\" (UID: \"e0f71887-ac32-4542-9a91-d1e350331dbc\") " pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.396928 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316c29a1-4346-4be0-be25-aa29a1a9dd0d-operator-scripts\") pod \"aodh-6a58-account-create-update-txbfr\" (UID: \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\") " pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.419922 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdqgb\" (UniqueName: \"kubernetes.io/projected/316c29a1-4346-4be0-be25-aa29a1a9dd0d-kube-api-access-jdqgb\") pod \"aodh-6a58-account-create-update-txbfr\" (UID: \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\") " pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.423022 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bf04a5a6-de08-4983-8986-c0c949090943","Type":"ContainerStarted","Data":"c63b6e6c01b11dcf2df1d4790bc24afdbae1f0cde40168fb4e95ee8fc277a5b0"} Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.423121 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bf04a5a6-de08-4983-8986-c0c949090943","Type":"ContainerStarted","Data":"2c9f943dd978740796ae5db72442514cf48fccc4a7c71296d92fb1ee9dc4fe45"} Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.426591 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7558z\" (UniqueName: \"kubernetes.io/projected/e0f71887-ac32-4542-9a91-d1e350331dbc-kube-api-access-7558z\") pod \"aodh-db-create-jljwh\" (UID: \"e0f71887-ac32-4542-9a91-d1e350331dbc\") " pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.446937 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.4469188920000002 podStartE2EDuration="2.446918892s" podCreationTimestamp="2025-12-05 09:05:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:05:05.441283079 +0000 UTC m=+2969.929947721" watchObservedRunningTime="2025-12-05 09:05:05.446918892 +0000 UTC m=+2969.935583504" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.464384 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.504867 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.919731 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jljwh"] Dec 05 09:05:05 crc kubenswrapper[4876]: I1205 09:05:05.968745 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-6a58-account-create-update-txbfr"] Dec 05 09:05:05 crc kubenswrapper[4876]: W1205 09:05:05.969060 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod316c29a1_4346_4be0_be25_aa29a1a9dd0d.slice/crio-6468052fb0e2e6bbc549deffcd04ef45f2776157128b93c97560e012fb617ed9 WatchSource:0}: Error finding container 6468052fb0e2e6bbc549deffcd04ef45f2776157128b93c97560e012fb617ed9: Status 404 returned error can't find the container with id 6468052fb0e2e6bbc549deffcd04ef45f2776157128b93c97560e012fb617ed9 Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.127713 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.130657 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="86ac21be-c7c2-4035-a32c-3f24715c3248" podUID="bf04a5a6-de08-4983-8986-c0c949090943" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.312724 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config-secret\") pod \"86ac21be-c7c2-4035-a32c-3f24715c3248\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.312865 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-combined-ca-bundle\") pod \"86ac21be-c7c2-4035-a32c-3f24715c3248\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.312957 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config\") pod \"86ac21be-c7c2-4035-a32c-3f24715c3248\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.313176 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sdxt\" (UniqueName: \"kubernetes.io/projected/86ac21be-c7c2-4035-a32c-3f24715c3248-kube-api-access-2sdxt\") pod \"86ac21be-c7c2-4035-a32c-3f24715c3248\" (UID: \"86ac21be-c7c2-4035-a32c-3f24715c3248\") " Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.323533 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ac21be-c7c2-4035-a32c-3f24715c3248-kube-api-access-2sdxt" (OuterVolumeSpecName: "kube-api-access-2sdxt") pod "86ac21be-c7c2-4035-a32c-3f24715c3248" (UID: "86ac21be-c7c2-4035-a32c-3f24715c3248"). InnerVolumeSpecName "kube-api-access-2sdxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.340489 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86ac21be-c7c2-4035-a32c-3f24715c3248" (UID: "86ac21be-c7c2-4035-a32c-3f24715c3248"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.342839 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "86ac21be-c7c2-4035-a32c-3f24715c3248" (UID: "86ac21be-c7c2-4035-a32c-3f24715c3248"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.365316 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "86ac21be-c7c2-4035-a32c-3f24715c3248" (UID: "86ac21be-c7c2-4035-a32c-3f24715c3248"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.416048 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sdxt\" (UniqueName: \"kubernetes.io/projected/86ac21be-c7c2-4035-a32c-3f24715c3248-kube-api-access-2sdxt\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.416084 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.416098 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86ac21be-c7c2-4035-a32c-3f24715c3248-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.416114 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/86ac21be-c7c2-4035-a32c-3f24715c3248-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.432254 4876 generic.go:334] "Generic (PLEG): container finished" podID="e0f71887-ac32-4542-9a91-d1e350331dbc" containerID="79e81b17f0124fd628a2e83af4ad9805c82f27c4cc5e189df2218ebe5c677100" exitCode=0 Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.432341 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jljwh" event={"ID":"e0f71887-ac32-4542-9a91-d1e350331dbc","Type":"ContainerDied","Data":"79e81b17f0124fd628a2e83af4ad9805c82f27c4cc5e189df2218ebe5c677100"} Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.432392 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jljwh" event={"ID":"e0f71887-ac32-4542-9a91-d1e350331dbc","Type":"ContainerStarted","Data":"af1c3728763a9b4f94ca240f9faa4fea42003c3dd0eace723203bcf799f33262"} Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.436020 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-6a58-account-create-update-txbfr" event={"ID":"316c29a1-4346-4be0-be25-aa29a1a9dd0d","Type":"ContainerStarted","Data":"ac3466d10336d513d8332d7902558d8e71c5e3bdb7bc711014e1eab9c5d57166"} Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.436061 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-6a58-account-create-update-txbfr" event={"ID":"316c29a1-4346-4be0-be25-aa29a1a9dd0d","Type":"ContainerStarted","Data":"6468052fb0e2e6bbc549deffcd04ef45f2776157128b93c97560e012fb617ed9"} Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.438551 4876 generic.go:334] "Generic (PLEG): container finished" podID="86ac21be-c7c2-4035-a32c-3f24715c3248" containerID="48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46" exitCode=137 Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.438677 4876 scope.go:117] "RemoveContainer" containerID="48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.438782 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.467680 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="86ac21be-c7c2-4035-a32c-3f24715c3248" podUID="bf04a5a6-de08-4983-8986-c0c949090943" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.469635 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-6a58-account-create-update-txbfr" podStartSLOduration=1.469615332 podStartE2EDuration="1.469615332s" podCreationTimestamp="2025-12-05 09:05:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:05:06.463448434 +0000 UTC m=+2970.952113056" watchObservedRunningTime="2025-12-05 09:05:06.469615332 +0000 UTC m=+2970.958279954" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.547698 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="86ac21be-c7c2-4035-a32c-3f24715c3248" podUID="bf04a5a6-de08-4983-8986-c0c949090943" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.555705 4876 scope.go:117] "RemoveContainer" containerID="48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46" Dec 05 09:05:06 crc kubenswrapper[4876]: E1205 09:05:06.556270 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46\": container with ID starting with 48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46 not found: ID does not exist" containerID="48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46" Dec 05 09:05:06 crc kubenswrapper[4876]: I1205 09:05:06.556310 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46"} err="failed to get container status \"48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46\": rpc error: code = NotFound desc = could not find container \"48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46\": container with ID starting with 48b321e4dc09624db9e47cd3a745aa20fd13ee9f753d3e41e53725b04ad58c46 not found: ID does not exist" Dec 05 09:05:07 crc kubenswrapper[4876]: I1205 09:05:07.449667 4876 generic.go:334] "Generic (PLEG): container finished" podID="316c29a1-4346-4be0-be25-aa29a1a9dd0d" containerID="ac3466d10336d513d8332d7902558d8e71c5e3bdb7bc711014e1eab9c5d57166" exitCode=0 Dec 05 09:05:07 crc kubenswrapper[4876]: I1205 09:05:07.449750 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-6a58-account-create-update-txbfr" event={"ID":"316c29a1-4346-4be0-be25-aa29a1a9dd0d","Type":"ContainerDied","Data":"ac3466d10336d513d8332d7902558d8e71c5e3bdb7bc711014e1eab9c5d57166"} Dec 05 09:05:07 crc kubenswrapper[4876]: I1205 09:05:07.804660 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:07 crc kubenswrapper[4876]: I1205 09:05:07.837956 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86ac21be-c7c2-4035-a32c-3f24715c3248" path="/var/lib/kubelet/pods/86ac21be-c7c2-4035-a32c-3f24715c3248/volumes" Dec 05 09:05:07 crc kubenswrapper[4876]: I1205 09:05:07.943035 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0f71887-ac32-4542-9a91-d1e350331dbc-operator-scripts\") pod \"e0f71887-ac32-4542-9a91-d1e350331dbc\" (UID: \"e0f71887-ac32-4542-9a91-d1e350331dbc\") " Dec 05 09:05:07 crc kubenswrapper[4876]: I1205 09:05:07.943355 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7558z\" (UniqueName: \"kubernetes.io/projected/e0f71887-ac32-4542-9a91-d1e350331dbc-kube-api-access-7558z\") pod \"e0f71887-ac32-4542-9a91-d1e350331dbc\" (UID: \"e0f71887-ac32-4542-9a91-d1e350331dbc\") " Dec 05 09:05:07 crc kubenswrapper[4876]: I1205 09:05:07.944051 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0f71887-ac32-4542-9a91-d1e350331dbc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e0f71887-ac32-4542-9a91-d1e350331dbc" (UID: "e0f71887-ac32-4542-9a91-d1e350331dbc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:05:07 crc kubenswrapper[4876]: I1205 09:05:07.948548 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f71887-ac32-4542-9a91-d1e350331dbc-kube-api-access-7558z" (OuterVolumeSpecName: "kube-api-access-7558z") pod "e0f71887-ac32-4542-9a91-d1e350331dbc" (UID: "e0f71887-ac32-4542-9a91-d1e350331dbc"). InnerVolumeSpecName "kube-api-access-7558z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.045464 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0f71887-ac32-4542-9a91-d1e350331dbc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.045503 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7558z\" (UniqueName: \"kubernetes.io/projected/e0f71887-ac32-4542-9a91-d1e350331dbc-kube-api-access-7558z\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.467099 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jljwh" Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.467182 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jljwh" event={"ID":"e0f71887-ac32-4542-9a91-d1e350331dbc","Type":"ContainerDied","Data":"af1c3728763a9b4f94ca240f9faa4fea42003c3dd0eace723203bcf799f33262"} Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.467939 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af1c3728763a9b4f94ca240f9faa4fea42003c3dd0eace723203bcf799f33262" Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.804319 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.960342 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdqgb\" (UniqueName: \"kubernetes.io/projected/316c29a1-4346-4be0-be25-aa29a1a9dd0d-kube-api-access-jdqgb\") pod \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\" (UID: \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\") " Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.960602 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316c29a1-4346-4be0-be25-aa29a1a9dd0d-operator-scripts\") pod \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\" (UID: \"316c29a1-4346-4be0-be25-aa29a1a9dd0d\") " Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.961291 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/316c29a1-4346-4be0-be25-aa29a1a9dd0d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "316c29a1-4346-4be0-be25-aa29a1a9dd0d" (UID: "316c29a1-4346-4be0-be25-aa29a1a9dd0d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:05:08 crc kubenswrapper[4876]: I1205 09:05:08.966369 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/316c29a1-4346-4be0-be25-aa29a1a9dd0d-kube-api-access-jdqgb" (OuterVolumeSpecName: "kube-api-access-jdqgb") pod "316c29a1-4346-4be0-be25-aa29a1a9dd0d" (UID: "316c29a1-4346-4be0-be25-aa29a1a9dd0d"). InnerVolumeSpecName "kube-api-access-jdqgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:09 crc kubenswrapper[4876]: I1205 09:05:09.063192 4876 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316c29a1-4346-4be0-be25-aa29a1a9dd0d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:09 crc kubenswrapper[4876]: I1205 09:05:09.063524 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdqgb\" (UniqueName: \"kubernetes.io/projected/316c29a1-4346-4be0-be25-aa29a1a9dd0d-kube-api-access-jdqgb\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:09 crc kubenswrapper[4876]: I1205 09:05:09.479712 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-6a58-account-create-update-txbfr" event={"ID":"316c29a1-4346-4be0-be25-aa29a1a9dd0d","Type":"ContainerDied","Data":"6468052fb0e2e6bbc549deffcd04ef45f2776157128b93c97560e012fb617ed9"} Dec 05 09:05:09 crc kubenswrapper[4876]: I1205 09:05:09.479758 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6468052fb0e2e6bbc549deffcd04ef45f2776157128b93c97560e012fb617ed9" Dec 05 09:05:09 crc kubenswrapper[4876]: I1205 09:05:09.479860 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-6a58-account-create-update-txbfr" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.572584 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-6wsbn"] Dec 05 09:05:10 crc kubenswrapper[4876]: E1205 09:05:10.573487 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316c29a1-4346-4be0-be25-aa29a1a9dd0d" containerName="mariadb-account-create-update" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.573521 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="316c29a1-4346-4be0-be25-aa29a1a9dd0d" containerName="mariadb-account-create-update" Dec 05 09:05:10 crc kubenswrapper[4876]: E1205 09:05:10.573547 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f71887-ac32-4542-9a91-d1e350331dbc" containerName="mariadb-database-create" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.573564 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f71887-ac32-4542-9a91-d1e350331dbc" containerName="mariadb-database-create" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.574196 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f71887-ac32-4542-9a91-d1e350331dbc" containerName="mariadb-database-create" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.574258 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="316c29a1-4346-4be0-be25-aa29a1a9dd0d" containerName="mariadb-account-create-update" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.575776 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.577746 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.578665 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.578683 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.579205 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tqf85" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.586526 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6wsbn"] Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.694017 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-config-data\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.694081 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-scripts\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.694127 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j8km\" (UniqueName: \"kubernetes.io/projected/52564344-c699-4703-bc3b-5170d5e891c7-kube-api-access-4j8km\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.694167 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-combined-ca-bundle\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.796298 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-config-data\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.796359 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-scripts\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.796406 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j8km\" (UniqueName: \"kubernetes.io/projected/52564344-c699-4703-bc3b-5170d5e891c7-kube-api-access-4j8km\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.796460 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-combined-ca-bundle\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.802401 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-combined-ca-bundle\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.804190 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-scripts\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.813419 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-config-data\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.818840 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j8km\" (UniqueName: \"kubernetes.io/projected/52564344-c699-4703-bc3b-5170d5e891c7-kube-api-access-4j8km\") pod \"aodh-db-sync-6wsbn\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:10 crc kubenswrapper[4876]: I1205 09:05:10.937030 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:11 crc kubenswrapper[4876]: I1205 09:05:11.389349 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-6wsbn"] Dec 05 09:05:11 crc kubenswrapper[4876]: W1205 09:05:11.402682 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52564344_c699_4703_bc3b_5170d5e891c7.slice/crio-1c5dbdad003d3bccd07fd7d84827539188ee64bdb6cc35bcd41e9a15c962c3e0 WatchSource:0}: Error finding container 1c5dbdad003d3bccd07fd7d84827539188ee64bdb6cc35bcd41e9a15c962c3e0: Status 404 returned error can't find the container with id 1c5dbdad003d3bccd07fd7d84827539188ee64bdb6cc35bcd41e9a15c962c3e0 Dec 05 09:05:11 crc kubenswrapper[4876]: I1205 09:05:11.496391 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6wsbn" event={"ID":"52564344-c699-4703-bc3b-5170d5e891c7","Type":"ContainerStarted","Data":"1c5dbdad003d3bccd07fd7d84827539188ee64bdb6cc35bcd41e9a15c962c3e0"} Dec 05 09:05:16 crc kubenswrapper[4876]: I1205 09:05:16.546350 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6wsbn" event={"ID":"52564344-c699-4703-bc3b-5170d5e891c7","Type":"ContainerStarted","Data":"e1af3f3a242a1506c4cdd1a68145e367df462320c7f7f1f1673d0f595b09dfa6"} Dec 05 09:05:16 crc kubenswrapper[4876]: I1205 09:05:16.582279 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-6wsbn" podStartSLOduration=2.400338033 podStartE2EDuration="6.582253095s" podCreationTimestamp="2025-12-05 09:05:10 +0000 UTC" firstStartedPulling="2025-12-05 09:05:11.407185442 +0000 UTC m=+2975.895850064" lastFinishedPulling="2025-12-05 09:05:15.589100504 +0000 UTC m=+2980.077765126" observedRunningTime="2025-12-05 09:05:16.561628735 +0000 UTC m=+2981.050293367" watchObservedRunningTime="2025-12-05 09:05:16.582253095 +0000 UTC m=+2981.070917717" Dec 05 09:05:17 crc kubenswrapper[4876]: I1205 09:05:17.824135 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:05:17 crc kubenswrapper[4876]: E1205 09:05:17.824757 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:05:18 crc kubenswrapper[4876]: I1205 09:05:18.566128 4876 generic.go:334] "Generic (PLEG): container finished" podID="52564344-c699-4703-bc3b-5170d5e891c7" containerID="e1af3f3a242a1506c4cdd1a68145e367df462320c7f7f1f1673d0f595b09dfa6" exitCode=0 Dec 05 09:05:18 crc kubenswrapper[4876]: I1205 09:05:18.566177 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6wsbn" event={"ID":"52564344-c699-4703-bc3b-5170d5e891c7","Type":"ContainerDied","Data":"e1af3f3a242a1506c4cdd1a68145e367df462320c7f7f1f1673d0f595b09dfa6"} Dec 05 09:05:19 crc kubenswrapper[4876]: I1205 09:05:19.908412 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.019302 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-combined-ca-bundle\") pod \"52564344-c699-4703-bc3b-5170d5e891c7\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.019396 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-scripts\") pod \"52564344-c699-4703-bc3b-5170d5e891c7\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.019497 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-config-data\") pod \"52564344-c699-4703-bc3b-5170d5e891c7\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.020286 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j8km\" (UniqueName: \"kubernetes.io/projected/52564344-c699-4703-bc3b-5170d5e891c7-kube-api-access-4j8km\") pod \"52564344-c699-4703-bc3b-5170d5e891c7\" (UID: \"52564344-c699-4703-bc3b-5170d5e891c7\") " Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.025360 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52564344-c699-4703-bc3b-5170d5e891c7-kube-api-access-4j8km" (OuterVolumeSpecName: "kube-api-access-4j8km") pod "52564344-c699-4703-bc3b-5170d5e891c7" (UID: "52564344-c699-4703-bc3b-5170d5e891c7"). InnerVolumeSpecName "kube-api-access-4j8km". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.029051 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-scripts" (OuterVolumeSpecName: "scripts") pod "52564344-c699-4703-bc3b-5170d5e891c7" (UID: "52564344-c699-4703-bc3b-5170d5e891c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.048763 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52564344-c699-4703-bc3b-5170d5e891c7" (UID: "52564344-c699-4703-bc3b-5170d5e891c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.066296 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-config-data" (OuterVolumeSpecName: "config-data") pod "52564344-c699-4703-bc3b-5170d5e891c7" (UID: "52564344-c699-4703-bc3b-5170d5e891c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.123218 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.123270 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.123296 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52564344-c699-4703-bc3b-5170d5e891c7-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.123314 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j8km\" (UniqueName: \"kubernetes.io/projected/52564344-c699-4703-bc3b-5170d5e891c7-kube-api-access-4j8km\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.583870 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-6wsbn" event={"ID":"52564344-c699-4703-bc3b-5170d5e891c7","Type":"ContainerDied","Data":"1c5dbdad003d3bccd07fd7d84827539188ee64bdb6cc35bcd41e9a15c962c3e0"} Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.583925 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c5dbdad003d3bccd07fd7d84827539188ee64bdb6cc35bcd41e9a15c962c3e0" Dec 05 09:05:20 crc kubenswrapper[4876]: I1205 09:05:20.583968 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-6wsbn" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.677104 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 05 09:05:25 crc kubenswrapper[4876]: E1205 09:05:25.677999 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52564344-c699-4703-bc3b-5170d5e891c7" containerName="aodh-db-sync" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.678013 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="52564344-c699-4703-bc3b-5170d5e891c7" containerName="aodh-db-sync" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.678228 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="52564344-c699-4703-bc3b-5170d5e891c7" containerName="aodh-db-sync" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.679809 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.681687 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.682254 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.691177 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.692552 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tqf85" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.835509 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-config-data\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.836108 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.836169 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-scripts\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.836248 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26gwr\" (UniqueName: \"kubernetes.io/projected/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-kube-api-access-26gwr\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.938548 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.938697 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-scripts\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.938762 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26gwr\" (UniqueName: \"kubernetes.io/projected/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-kube-api-access-26gwr\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.938844 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-config-data\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.944249 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.944307 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-scripts\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.944393 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-config-data\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:25 crc kubenswrapper[4876]: I1205 09:05:25.961024 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26gwr\" (UniqueName: \"kubernetes.io/projected/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-kube-api-access-26gwr\") pod \"aodh-0\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " pod="openstack/aodh-0" Dec 05 09:05:26 crc kubenswrapper[4876]: I1205 09:05:26.010744 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:05:26 crc kubenswrapper[4876]: I1205 09:05:26.507252 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:05:26 crc kubenswrapper[4876]: I1205 09:05:26.635665 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerStarted","Data":"dd5b72eeca71d6f37abdf8320f66dbb42e587b983ffd8307f4c6d079ecf7aeea"} Dec 05 09:05:27 crc kubenswrapper[4876]: I1205 09:05:27.644742 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerStarted","Data":"58c40123f27cec41a973a95e697e729ba51afcfae2f84a465b4d698f255b640b"} Dec 05 09:05:27 crc kubenswrapper[4876]: I1205 09:05:27.730224 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:27 crc kubenswrapper[4876]: I1205 09:05:27.732690 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="ceilometer-central-agent" containerID="cri-o://617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8" gracePeriod=30 Dec 05 09:05:27 crc kubenswrapper[4876]: I1205 09:05:27.732977 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="proxy-httpd" containerID="cri-o://e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3" gracePeriod=30 Dec 05 09:05:27 crc kubenswrapper[4876]: I1205 09:05:27.733095 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="sg-core" containerID="cri-o://68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765" gracePeriod=30 Dec 05 09:05:27 crc kubenswrapper[4876]: I1205 09:05:27.733182 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="ceilometer-notification-agent" containerID="cri-o://63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326" gracePeriod=30 Dec 05 09:05:28 crc kubenswrapper[4876]: I1205 09:05:28.467815 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 05 09:05:28 crc kubenswrapper[4876]: I1205 09:05:28.656420 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerStarted","Data":"fcc50679dc5e7bf9bdc96a33cdf225d6affd8d968decd8d8176981bdda6b72ea"} Dec 05 09:05:28 crc kubenswrapper[4876]: I1205 09:05:28.660975 4876 generic.go:334] "Generic (PLEG): container finished" podID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerID="e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3" exitCode=0 Dec 05 09:05:28 crc kubenswrapper[4876]: I1205 09:05:28.660999 4876 generic.go:334] "Generic (PLEG): container finished" podID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerID="68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765" exitCode=2 Dec 05 09:05:28 crc kubenswrapper[4876]: I1205 09:05:28.661006 4876 generic.go:334] "Generic (PLEG): container finished" podID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerID="617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8" exitCode=0 Dec 05 09:05:28 crc kubenswrapper[4876]: I1205 09:05:28.661019 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerDied","Data":"e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3"} Dec 05 09:05:28 crc kubenswrapper[4876]: I1205 09:05:28.661034 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerDied","Data":"68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765"} Dec 05 09:05:28 crc kubenswrapper[4876]: I1205 09:05:28.661043 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerDied","Data":"617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8"} Dec 05 09:05:30 crc kubenswrapper[4876]: I1205 09:05:30.684066 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerStarted","Data":"1a8744b9cf6c51a72d7c96b9905a206e7717bb53696de40f5d075ad7d47b0149"} Dec 05 09:05:31 crc kubenswrapper[4876]: I1205 09:05:31.693877 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerStarted","Data":"3f9708ae3f21a07a669884e0e593c4ba0266da6622b0878dfb4a7296efb95c9d"} Dec 05 09:05:31 crc kubenswrapper[4876]: I1205 09:05:31.694251 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-api" containerID="cri-o://58c40123f27cec41a973a95e697e729ba51afcfae2f84a465b4d698f255b640b" gracePeriod=30 Dec 05 09:05:31 crc kubenswrapper[4876]: I1205 09:05:31.694313 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-listener" containerID="cri-o://3f9708ae3f21a07a669884e0e593c4ba0266da6622b0878dfb4a7296efb95c9d" gracePeriod=30 Dec 05 09:05:31 crc kubenswrapper[4876]: I1205 09:05:31.694362 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-notifier" containerID="cri-o://1a8744b9cf6c51a72d7c96b9905a206e7717bb53696de40f5d075ad7d47b0149" gracePeriod=30 Dec 05 09:05:31 crc kubenswrapper[4876]: I1205 09:05:31.694409 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-evaluator" containerID="cri-o://fcc50679dc5e7bf9bdc96a33cdf225d6affd8d968decd8d8176981bdda6b72ea" gracePeriod=30 Dec 05 09:05:31 crc kubenswrapper[4876]: I1205 09:05:31.716691 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.9349458510000002 podStartE2EDuration="6.716671162s" podCreationTimestamp="2025-12-05 09:05:25 +0000 UTC" firstStartedPulling="2025-12-05 09:05:26.527623322 +0000 UTC m=+2991.016287944" lastFinishedPulling="2025-12-05 09:05:31.309348633 +0000 UTC m=+2995.798013255" observedRunningTime="2025-12-05 09:05:31.713405395 +0000 UTC m=+2996.202070027" watchObservedRunningTime="2025-12-05 09:05:31.716671162 +0000 UTC m=+2996.205335784" Dec 05 09:05:32 crc kubenswrapper[4876]: I1205 09:05:32.704494 4876 generic.go:334] "Generic (PLEG): container finished" podID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerID="fcc50679dc5e7bf9bdc96a33cdf225d6affd8d968decd8d8176981bdda6b72ea" exitCode=0 Dec 05 09:05:32 crc kubenswrapper[4876]: I1205 09:05:32.704795 4876 generic.go:334] "Generic (PLEG): container finished" podID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerID="58c40123f27cec41a973a95e697e729ba51afcfae2f84a465b4d698f255b640b" exitCode=0 Dec 05 09:05:32 crc kubenswrapper[4876]: I1205 09:05:32.704576 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerDied","Data":"fcc50679dc5e7bf9bdc96a33cdf225d6affd8d968decd8d8176981bdda6b72ea"} Dec 05 09:05:32 crc kubenswrapper[4876]: I1205 09:05:32.704837 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerDied","Data":"58c40123f27cec41a973a95e697e729ba51afcfae2f84a465b4d698f255b640b"} Dec 05 09:05:32 crc kubenswrapper[4876]: I1205 09:05:32.824579 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:05:32 crc kubenswrapper[4876]: E1205 09:05:32.825568 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.120998 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.196682 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-log-httpd\") pod \"d065d71e-e44d-426d-87e9-5924a1ed5784\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.196724 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-run-httpd\") pod \"d065d71e-e44d-426d-87e9-5924a1ed5784\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.196811 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-scripts\") pod \"d065d71e-e44d-426d-87e9-5924a1ed5784\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.196838 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-sg-core-conf-yaml\") pod \"d065d71e-e44d-426d-87e9-5924a1ed5784\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.196886 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-config-data\") pod \"d065d71e-e44d-426d-87e9-5924a1ed5784\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.196978 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z66cf\" (UniqueName: \"kubernetes.io/projected/d065d71e-e44d-426d-87e9-5924a1ed5784-kube-api-access-z66cf\") pod \"d065d71e-e44d-426d-87e9-5924a1ed5784\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.197007 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-combined-ca-bundle\") pod \"d065d71e-e44d-426d-87e9-5924a1ed5784\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.197051 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-ceilometer-tls-certs\") pod \"d065d71e-e44d-426d-87e9-5924a1ed5784\" (UID: \"d065d71e-e44d-426d-87e9-5924a1ed5784\") " Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.200160 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d065d71e-e44d-426d-87e9-5924a1ed5784" (UID: "d065d71e-e44d-426d-87e9-5924a1ed5784"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.200447 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d065d71e-e44d-426d-87e9-5924a1ed5784" (UID: "d065d71e-e44d-426d-87e9-5924a1ed5784"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.214145 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d065d71e-e44d-426d-87e9-5924a1ed5784-kube-api-access-z66cf" (OuterVolumeSpecName: "kube-api-access-z66cf") pod "d065d71e-e44d-426d-87e9-5924a1ed5784" (UID: "d065d71e-e44d-426d-87e9-5924a1ed5784"). InnerVolumeSpecName "kube-api-access-z66cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.214236 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-scripts" (OuterVolumeSpecName: "scripts") pod "d065d71e-e44d-426d-87e9-5924a1ed5784" (UID: "d065d71e-e44d-426d-87e9-5924a1ed5784"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.237628 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d065d71e-e44d-426d-87e9-5924a1ed5784" (UID: "d065d71e-e44d-426d-87e9-5924a1ed5784"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.262168 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d065d71e-e44d-426d-87e9-5924a1ed5784" (UID: "d065d71e-e44d-426d-87e9-5924a1ed5784"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.285306 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d065d71e-e44d-426d-87e9-5924a1ed5784" (UID: "d065d71e-e44d-426d-87e9-5924a1ed5784"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.300642 4876 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.300685 4876 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d065d71e-e44d-426d-87e9-5924a1ed5784-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.300696 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.300711 4876 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.300748 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z66cf\" (UniqueName: \"kubernetes.io/projected/d065d71e-e44d-426d-87e9-5924a1ed5784-kube-api-access-z66cf\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.300760 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.300769 4876 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.304397 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-config-data" (OuterVolumeSpecName: "config-data") pod "d065d71e-e44d-426d-87e9-5924a1ed5784" (UID: "d065d71e-e44d-426d-87e9-5924a1ed5784"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.403417 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d065d71e-e44d-426d-87e9-5924a1ed5784-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.714638 4876 generic.go:334] "Generic (PLEG): container finished" podID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerID="63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326" exitCode=0 Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.714781 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerDied","Data":"63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326"} Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.715024 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d065d71e-e44d-426d-87e9-5924a1ed5784","Type":"ContainerDied","Data":"39c72c126e9850f6c76e9e6df2f39cd7063bd3f7a876509778d56064d237cb9f"} Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.715046 4876 scope.go:117] "RemoveContainer" containerID="e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.714915 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.745193 4876 scope.go:117] "RemoveContainer" containerID="68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.757567 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.775451 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.787993 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:33 crc kubenswrapper[4876]: E1205 09:05:33.788590 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="sg-core" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.788617 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="sg-core" Dec 05 09:05:33 crc kubenswrapper[4876]: E1205 09:05:33.788641 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="ceilometer-notification-agent" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.788653 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="ceilometer-notification-agent" Dec 05 09:05:33 crc kubenswrapper[4876]: E1205 09:05:33.788694 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="ceilometer-central-agent" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.788707 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="ceilometer-central-agent" Dec 05 09:05:33 crc kubenswrapper[4876]: E1205 09:05:33.788727 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="proxy-httpd" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.788737 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="proxy-httpd" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.789017 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="ceilometer-notification-agent" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.789044 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="proxy-httpd" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.789057 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="ceilometer-central-agent" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.789075 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" containerName="sg-core" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.791121 4876 scope.go:117] "RemoveContainer" containerID="63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.791581 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.793368 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.794081 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.794305 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.798873 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.840429 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d065d71e-e44d-426d-87e9-5924a1ed5784" path="/var/lib/kubelet/pods/d065d71e-e44d-426d-87e9-5924a1ed5784/volumes" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.853445 4876 scope.go:117] "RemoveContainer" containerID="617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.878844 4876 scope.go:117] "RemoveContainer" containerID="e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3" Dec 05 09:05:33 crc kubenswrapper[4876]: E1205 09:05:33.879342 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3\": container with ID starting with e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3 not found: ID does not exist" containerID="e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.879403 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3"} err="failed to get container status \"e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3\": rpc error: code = NotFound desc = could not find container \"e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3\": container with ID starting with e9012c349a88f5c2a153f3eb9cfc5516c15686702329d82fd6856885dd751eb3 not found: ID does not exist" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.879434 4876 scope.go:117] "RemoveContainer" containerID="68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765" Dec 05 09:05:33 crc kubenswrapper[4876]: E1205 09:05:33.880132 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765\": container with ID starting with 68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765 not found: ID does not exist" containerID="68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.880167 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765"} err="failed to get container status \"68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765\": rpc error: code = NotFound desc = could not find container \"68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765\": container with ID starting with 68288bba98ac868cfb275e42e8bbfe46691ed65805ae33fe026451e6bea41765 not found: ID does not exist" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.880189 4876 scope.go:117] "RemoveContainer" containerID="63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326" Dec 05 09:05:33 crc kubenswrapper[4876]: E1205 09:05:33.880522 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326\": container with ID starting with 63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326 not found: ID does not exist" containerID="63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.880549 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326"} err="failed to get container status \"63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326\": rpc error: code = NotFound desc = could not find container \"63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326\": container with ID starting with 63ac11a4c5fe2b753ff0d6b3917afb7c12b098d34a0cd8898bfccde8b0315326 not found: ID does not exist" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.880563 4876 scope.go:117] "RemoveContainer" containerID="617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8" Dec 05 09:05:33 crc kubenswrapper[4876]: E1205 09:05:33.880817 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8\": container with ID starting with 617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8 not found: ID does not exist" containerID="617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.880858 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8"} err="failed to get container status \"617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8\": rpc error: code = NotFound desc = could not find container \"617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8\": container with ID starting with 617e20fbb83f27ad997bc30e0f53340bf249e6eb14f5ad15239936a302d6c4f8 not found: ID does not exist" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.918809 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2143f266-92fd-4c74-b305-0e1d95d3044b-log-httpd\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.918932 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-config-data\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.918974 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.919006 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmq4j\" (UniqueName: \"kubernetes.io/projected/2143f266-92fd-4c74-b305-0e1d95d3044b-kube-api-access-nmq4j\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.919045 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-scripts\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.919060 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.919202 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2143f266-92fd-4c74-b305-0e1d95d3044b-run-httpd\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:33 crc kubenswrapper[4876]: I1205 09:05:33.919374 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.021052 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2143f266-92fd-4c74-b305-0e1d95d3044b-log-httpd\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.021114 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-config-data\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.021140 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.021165 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmq4j\" (UniqueName: \"kubernetes.io/projected/2143f266-92fd-4c74-b305-0e1d95d3044b-kube-api-access-nmq4j\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.021207 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-scripts\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.021221 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.021249 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2143f266-92fd-4c74-b305-0e1d95d3044b-run-httpd\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.021316 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.022728 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2143f266-92fd-4c74-b305-0e1d95d3044b-log-httpd\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.023002 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2143f266-92fd-4c74-b305-0e1d95d3044b-run-httpd\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.026959 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.027573 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.029195 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-config-data\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.033444 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-scripts\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.034329 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2143f266-92fd-4c74-b305-0e1d95d3044b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.047476 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmq4j\" (UniqueName: \"kubernetes.io/projected/2143f266-92fd-4c74-b305-0e1d95d3044b-kube-api-access-nmq4j\") pod \"ceilometer-0\" (UID: \"2143f266-92fd-4c74-b305-0e1d95d3044b\") " pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.163465 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:05:34 crc kubenswrapper[4876]: W1205 09:05:34.636524 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2143f266_92fd_4c74_b305_0e1d95d3044b.slice/crio-bbe21aafd57b6360e29da2694f35c4ba7779241c962f57c9e7e53924063bc371 WatchSource:0}: Error finding container bbe21aafd57b6360e29da2694f35c4ba7779241c962f57c9e7e53924063bc371: Status 404 returned error can't find the container with id bbe21aafd57b6360e29da2694f35c4ba7779241c962f57c9e7e53924063bc371 Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.638309 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:05:34 crc kubenswrapper[4876]: I1205 09:05:34.727173 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2143f266-92fd-4c74-b305-0e1d95d3044b","Type":"ContainerStarted","Data":"bbe21aafd57b6360e29da2694f35c4ba7779241c962f57c9e7e53924063bc371"} Dec 05 09:05:35 crc kubenswrapper[4876]: I1205 09:05:35.737961 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2143f266-92fd-4c74-b305-0e1d95d3044b","Type":"ContainerStarted","Data":"ce731e031f5c1de620f9e516a3cbadd295330d7ebcf81f66c97d5621ef9ced02"} Dec 05 09:05:36 crc kubenswrapper[4876]: I1205 09:05:36.749000 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2143f266-92fd-4c74-b305-0e1d95d3044b","Type":"ContainerStarted","Data":"48be26a76d53596b2c461d4431d9d6e3e6d796be587c7f582a8163e5a1b0f5c9"} Dec 05 09:05:37 crc kubenswrapper[4876]: I1205 09:05:37.760418 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2143f266-92fd-4c74-b305-0e1d95d3044b","Type":"ContainerStarted","Data":"e8ea71ce7d7baa52d45ffd657c4c117f71c9d54a827fa7a64af0ac70182ee7b2"} Dec 05 09:05:38 crc kubenswrapper[4876]: I1205 09:05:38.770050 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2143f266-92fd-4c74-b305-0e1d95d3044b","Type":"ContainerStarted","Data":"c892367a7d835895213febe3160cfbc4f0bf65bb80a29abf10c1900bb1bd6530"} Dec 05 09:05:38 crc kubenswrapper[4876]: I1205 09:05:38.770527 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:05:38 crc kubenswrapper[4876]: I1205 09:05:38.804271 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.395957442 podStartE2EDuration="5.804253436s" podCreationTimestamp="2025-12-05 09:05:33 +0000 UTC" firstStartedPulling="2025-12-05 09:05:34.64037377 +0000 UTC m=+2999.129038392" lastFinishedPulling="2025-12-05 09:05:38.048669764 +0000 UTC m=+3002.537334386" observedRunningTime="2025-12-05 09:05:38.797059625 +0000 UTC m=+3003.285724267" watchObservedRunningTime="2025-12-05 09:05:38.804253436 +0000 UTC m=+3003.292918058" Dec 05 09:05:47 crc kubenswrapper[4876]: I1205 09:05:47.824431 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:05:47 crc kubenswrapper[4876]: E1205 09:05:47.825634 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:06:00 crc kubenswrapper[4876]: I1205 09:06:00.824838 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:06:00 crc kubenswrapper[4876]: E1205 09:06:00.825826 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.038741 4876 generic.go:334] "Generic (PLEG): container finished" podID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerID="3f9708ae3f21a07a669884e0e593c4ba0266da6622b0878dfb4a7296efb95c9d" exitCode=137 Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.039166 4876 generic.go:334] "Generic (PLEG): container finished" podID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerID="1a8744b9cf6c51a72d7c96b9905a206e7717bb53696de40f5d075ad7d47b0149" exitCode=137 Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.038816 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerDied","Data":"3f9708ae3f21a07a669884e0e593c4ba0266da6622b0878dfb4a7296efb95c9d"} Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.039260 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerDied","Data":"1a8744b9cf6c51a72d7c96b9905a206e7717bb53696de40f5d075ad7d47b0149"} Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.192840 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.329719 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-scripts\") pod \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.330232 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26gwr\" (UniqueName: \"kubernetes.io/projected/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-kube-api-access-26gwr\") pod \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.330361 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-combined-ca-bundle\") pod \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.330399 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-config-data\") pod \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\" (UID: \"5c072d75-6af9-4076-8a23-a0e9c9ab2d71\") " Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.336289 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-kube-api-access-26gwr" (OuterVolumeSpecName: "kube-api-access-26gwr") pod "5c072d75-6af9-4076-8a23-a0e9c9ab2d71" (UID: "5c072d75-6af9-4076-8a23-a0e9c9ab2d71"). InnerVolumeSpecName "kube-api-access-26gwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.349158 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-scripts" (OuterVolumeSpecName: "scripts") pod "5c072d75-6af9-4076-8a23-a0e9c9ab2d71" (UID: "5c072d75-6af9-4076-8a23-a0e9c9ab2d71"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.433724 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26gwr\" (UniqueName: \"kubernetes.io/projected/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-kube-api-access-26gwr\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.433764 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.435729 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-config-data" (OuterVolumeSpecName: "config-data") pod "5c072d75-6af9-4076-8a23-a0e9c9ab2d71" (UID: "5c072d75-6af9-4076-8a23-a0e9c9ab2d71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.439957 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c072d75-6af9-4076-8a23-a0e9c9ab2d71" (UID: "5c072d75-6af9-4076-8a23-a0e9c9ab2d71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.535141 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:02 crc kubenswrapper[4876]: I1205 09:06:02.535400 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c072d75-6af9-4076-8a23-a0e9c9ab2d71-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.051172 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5c072d75-6af9-4076-8a23-a0e9c9ab2d71","Type":"ContainerDied","Data":"dd5b72eeca71d6f37abdf8320f66dbb42e587b983ffd8307f4c6d079ecf7aeea"} Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.051222 4876 scope.go:117] "RemoveContainer" containerID="3f9708ae3f21a07a669884e0e593c4ba0266da6622b0878dfb4a7296efb95c9d" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.051429 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.077790 4876 scope.go:117] "RemoveContainer" containerID="1a8744b9cf6c51a72d7c96b9905a206e7717bb53696de40f5d075ad7d47b0149" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.088065 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.099016 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.115449 4876 scope.go:117] "RemoveContainer" containerID="fcc50679dc5e7bf9bdc96a33cdf225d6affd8d968decd8d8176981bdda6b72ea" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.120198 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 05 09:06:03 crc kubenswrapper[4876]: E1205 09:06:03.120669 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-evaluator" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.120696 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-evaluator" Dec 05 09:06:03 crc kubenswrapper[4876]: E1205 09:06:03.120714 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-api" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.120722 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-api" Dec 05 09:06:03 crc kubenswrapper[4876]: E1205 09:06:03.120752 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-notifier" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.120769 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-notifier" Dec 05 09:06:03 crc kubenswrapper[4876]: E1205 09:06:03.120788 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-listener" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.120796 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-listener" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.121044 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-api" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.121075 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-notifier" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.121089 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-evaluator" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.121105 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" containerName="aodh-listener" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.123175 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.126129 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.126597 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.126868 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.127075 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.127430 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tqf85" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.129804 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.140060 4876 scope.go:117] "RemoveContainer" containerID="58c40123f27cec41a973a95e697e729ba51afcfae2f84a465b4d698f255b640b" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.247278 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.247704 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-config-data\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.247921 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfmxl\" (UniqueName: \"kubernetes.io/projected/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-kube-api-access-rfmxl\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.248185 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-scripts\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.248275 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-internal-tls-certs\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.248424 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-public-tls-certs\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.350039 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-scripts\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.350107 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-internal-tls-certs\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.350166 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-public-tls-certs\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.350231 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.350298 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-config-data\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.350353 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfmxl\" (UniqueName: \"kubernetes.io/projected/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-kube-api-access-rfmxl\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.355169 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-scripts\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.357380 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-public-tls-certs\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.364659 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-combined-ca-bundle\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.367493 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-internal-tls-certs\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.369631 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-config-data\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.377505 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfmxl\" (UniqueName: \"kubernetes.io/projected/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-kube-api-access-rfmxl\") pod \"aodh-0\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.439958 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.835802 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c072d75-6af9-4076-8a23-a0e9c9ab2d71" path="/var/lib/kubelet/pods/5c072d75-6af9-4076-8a23-a0e9c9ab2d71/volumes" Dec 05 09:06:03 crc kubenswrapper[4876]: I1205 09:06:03.888618 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:06:04 crc kubenswrapper[4876]: I1205 09:06:04.060660 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerStarted","Data":"2437d7ccb64300e95892f4fa2fbe5372a072e9a66bf962385fc428da8573a8a2"} Dec 05 09:06:04 crc kubenswrapper[4876]: I1205 09:06:04.180169 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 09:06:05 crc kubenswrapper[4876]: I1205 09:06:05.084122 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerStarted","Data":"17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872"} Dec 05 09:06:06 crc kubenswrapper[4876]: I1205 09:06:06.102042 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerStarted","Data":"445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c"} Dec 05 09:06:06 crc kubenswrapper[4876]: I1205 09:06:06.102401 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerStarted","Data":"01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e"} Dec 05 09:06:07 crc kubenswrapper[4876]: I1205 09:06:07.130012 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerStarted","Data":"7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15"} Dec 05 09:06:07 crc kubenswrapper[4876]: I1205 09:06:07.150109 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.566021683 podStartE2EDuration="4.150087466s" podCreationTimestamp="2025-12-05 09:06:03 +0000 UTC" firstStartedPulling="2025-12-05 09:06:03.893251366 +0000 UTC m=+3028.381915988" lastFinishedPulling="2025-12-05 09:06:06.477317159 +0000 UTC m=+3030.965981771" observedRunningTime="2025-12-05 09:06:07.147807506 +0000 UTC m=+3031.636472138" watchObservedRunningTime="2025-12-05 09:06:07.150087466 +0000 UTC m=+3031.638752108" Dec 05 09:06:12 crc kubenswrapper[4876]: I1205 09:06:12.823660 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:06:12 crc kubenswrapper[4876]: E1205 09:06:12.824360 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:06:24 crc kubenswrapper[4876]: I1205 09:06:24.824463 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:06:24 crc kubenswrapper[4876]: E1205 09:06:24.825732 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:06:35 crc kubenswrapper[4876]: I1205 09:06:35.834522 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:06:35 crc kubenswrapper[4876]: E1205 09:06:35.835165 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:06:50 crc kubenswrapper[4876]: I1205 09:06:50.824348 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:06:50 crc kubenswrapper[4876]: E1205 09:06:50.825375 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:07:04 crc kubenswrapper[4876]: I1205 09:07:04.823822 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:07:04 crc kubenswrapper[4876]: E1205 09:07:04.824457 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:07:17 crc kubenswrapper[4876]: I1205 09:07:17.824073 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:07:17 crc kubenswrapper[4876]: E1205 09:07:17.824768 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:07:32 crc kubenswrapper[4876]: I1205 09:07:32.827181 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:07:32 crc kubenswrapper[4876]: E1205 09:07:32.828117 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:07:43 crc kubenswrapper[4876]: I1205 09:07:43.824510 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:07:43 crc kubenswrapper[4876]: E1205 09:07:43.825456 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:07:57 crc kubenswrapper[4876]: I1205 09:07:57.824224 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:07:57 crc kubenswrapper[4876]: E1205 09:07:57.826552 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:08:11 crc kubenswrapper[4876]: I1205 09:08:11.824309 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:08:11 crc kubenswrapper[4876]: E1205 09:08:11.825587 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:08:25 crc kubenswrapper[4876]: I1205 09:08:25.831671 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:08:25 crc kubenswrapper[4876]: E1205 09:08:25.835931 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:08:38 crc kubenswrapper[4876]: I1205 09:08:38.825527 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:08:38 crc kubenswrapper[4876]: E1205 09:08:38.826501 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:08:50 crc kubenswrapper[4876]: I1205 09:08:50.824657 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:08:50 crc kubenswrapper[4876]: E1205 09:08:50.826008 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:09:04 crc kubenswrapper[4876]: I1205 09:09:04.825016 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:09:04 crc kubenswrapper[4876]: E1205 09:09:04.826024 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:09:05 crc kubenswrapper[4876]: I1205 09:09:05.916351 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/manager/0.log" Dec 05 09:09:16 crc kubenswrapper[4876]: I1205 09:09:16.824007 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:09:18 crc kubenswrapper[4876]: I1205 09:09:18.008477 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"df4e6443301eadd3c41cab90e72814aa0d48c3c05d4aabdd9044ffdf495e776d"} Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.280446 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv"] Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.284212 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.287253 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.310466 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv"] Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.424852 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.424993 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.425053 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df2z5\" (UniqueName: \"kubernetes.io/projected/1ee6bebd-088e-47ce-97a1-78d3132184f4-kube-api-access-df2z5\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.526468 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.526550 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df2z5\" (UniqueName: \"kubernetes.io/projected/1ee6bebd-088e-47ce-97a1-78d3132184f4-kube-api-access-df2z5\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.526721 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.527485 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.527601 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.550623 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df2z5\" (UniqueName: \"kubernetes.io/projected/1ee6bebd-088e-47ce-97a1-78d3132184f4-kube-api-access-df2z5\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:22 crc kubenswrapper[4876]: I1205 09:09:22.611769 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:23 crc kubenswrapper[4876]: I1205 09:09:23.099716 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv"] Dec 05 09:09:24 crc kubenswrapper[4876]: I1205 09:09:24.090614 4876 generic.go:334] "Generic (PLEG): container finished" podID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerID="1d4557f64dd9f8280391752939a36fbc66c982bf1c1251ed35f396de24cffc85" exitCode=0 Dec 05 09:09:24 crc kubenswrapper[4876]: I1205 09:09:24.090695 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" event={"ID":"1ee6bebd-088e-47ce-97a1-78d3132184f4","Type":"ContainerDied","Data":"1d4557f64dd9f8280391752939a36fbc66c982bf1c1251ed35f396de24cffc85"} Dec 05 09:09:24 crc kubenswrapper[4876]: I1205 09:09:24.091004 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" event={"ID":"1ee6bebd-088e-47ce-97a1-78d3132184f4","Type":"ContainerStarted","Data":"85b4435a6903e2415b895302a132d4c3b1893b901cb070088dd721e83136ccae"} Dec 05 09:09:24 crc kubenswrapper[4876]: I1205 09:09:24.093092 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:09:26 crc kubenswrapper[4876]: I1205 09:09:26.118372 4876 generic.go:334] "Generic (PLEG): container finished" podID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerID="66059b58bdf788468ce7a69757504d761096b93a0f41ee62b66e160f7ffe9cb0" exitCode=0 Dec 05 09:09:26 crc kubenswrapper[4876]: I1205 09:09:26.118474 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" event={"ID":"1ee6bebd-088e-47ce-97a1-78d3132184f4","Type":"ContainerDied","Data":"66059b58bdf788468ce7a69757504d761096b93a0f41ee62b66e160f7ffe9cb0"} Dec 05 09:09:27 crc kubenswrapper[4876]: I1205 09:09:27.133412 4876 generic.go:334] "Generic (PLEG): container finished" podID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerID="3515413a3e720f9225ea16b72d47991d3d7f8f3efa6eadf93d56c8db8b7960b6" exitCode=0 Dec 05 09:09:27 crc kubenswrapper[4876]: I1205 09:09:27.133558 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" event={"ID":"1ee6bebd-088e-47ce-97a1-78d3132184f4","Type":"ContainerDied","Data":"3515413a3e720f9225ea16b72d47991d3d7f8f3efa6eadf93d56c8db8b7960b6"} Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.496781 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.683434 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-bundle\") pod \"1ee6bebd-088e-47ce-97a1-78d3132184f4\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.683810 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-util\") pod \"1ee6bebd-088e-47ce-97a1-78d3132184f4\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.683958 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df2z5\" (UniqueName: \"kubernetes.io/projected/1ee6bebd-088e-47ce-97a1-78d3132184f4-kube-api-access-df2z5\") pod \"1ee6bebd-088e-47ce-97a1-78d3132184f4\" (UID: \"1ee6bebd-088e-47ce-97a1-78d3132184f4\") " Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.686767 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-bundle" (OuterVolumeSpecName: "bundle") pod "1ee6bebd-088e-47ce-97a1-78d3132184f4" (UID: "1ee6bebd-088e-47ce-97a1-78d3132184f4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.690056 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ee6bebd-088e-47ce-97a1-78d3132184f4-kube-api-access-df2z5" (OuterVolumeSpecName: "kube-api-access-df2z5") pod "1ee6bebd-088e-47ce-97a1-78d3132184f4" (UID: "1ee6bebd-088e-47ce-97a1-78d3132184f4"). InnerVolumeSpecName "kube-api-access-df2z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.787282 4876 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.787666 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df2z5\" (UniqueName: \"kubernetes.io/projected/1ee6bebd-088e-47ce-97a1-78d3132184f4-kube-api-access-df2z5\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.811489 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-util" (OuterVolumeSpecName: "util") pod "1ee6bebd-088e-47ce-97a1-78d3132184f4" (UID: "1ee6bebd-088e-47ce-97a1-78d3132184f4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:09:28 crc kubenswrapper[4876]: I1205 09:09:28.889274 4876 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ee6bebd-088e-47ce-97a1-78d3132184f4-util\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:29 crc kubenswrapper[4876]: I1205 09:09:29.156814 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" event={"ID":"1ee6bebd-088e-47ce-97a1-78d3132184f4","Type":"ContainerDied","Data":"85b4435a6903e2415b895302a132d4c3b1893b901cb070088dd721e83136ccae"} Dec 05 09:09:29 crc kubenswrapper[4876]: I1205 09:09:29.156872 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85b4435a6903e2415b895302a132d4c3b1893b901cb070088dd721e83136ccae" Dec 05 09:09:29 crc kubenswrapper[4876]: I1205 09:09:29.156958 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.203095 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9"] Dec 05 09:09:40 crc kubenswrapper[4876]: E1205 09:09:40.204206 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerName="extract" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.204224 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerName="extract" Dec 05 09:09:40 crc kubenswrapper[4876]: E1205 09:09:40.204242 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerName="pull" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.204249 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerName="pull" Dec 05 09:09:40 crc kubenswrapper[4876]: E1205 09:09:40.204259 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerName="util" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.204267 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerName="util" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.204550 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ee6bebd-088e-47ce-97a1-78d3132184f4" containerName="extract" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.205362 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.210459 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.211118 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-z276s" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.212945 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.216696 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.218215 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.222351 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.224012 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-hr57n" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.225689 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.255184 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.256721 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.267932 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.292140 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.350409 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cc522201-89c0-478e-8e63-4962e162b4e6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-c54zb\" (UID: \"cc522201-89c0-478e-8e63-4962e162b4e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.350543 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm5vk\" (UniqueName: \"kubernetes.io/projected/8c2118f9-ee5c-4fb7-a734-f9fbc098c53b-kube-api-access-nm5vk\") pod \"obo-prometheus-operator-668cf9dfbb-k9rt9\" (UID: \"8c2118f9-ee5c-4fb7-a734-f9fbc098c53b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.350602 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84bb503d-0098-4342-bad3-6dc8e2ae325b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-4vxch\" (UID: \"84bb503d-0098-4342-bad3-6dc8e2ae325b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.350690 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cc522201-89c0-478e-8e63-4962e162b4e6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-c54zb\" (UID: \"cc522201-89c0-478e-8e63-4962e162b4e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.350727 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84bb503d-0098-4342-bad3-6dc8e2ae325b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-4vxch\" (UID: \"84bb503d-0098-4342-bad3-6dc8e2ae325b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.370189 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-v7c8l"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.371878 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.375395 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-lqkbz" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.375565 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.381607 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-v7c8l"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.452954 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf4xd\" (UniqueName: \"kubernetes.io/projected/714c939d-43ee-4564-851d-198cc78fbf7f-kube-api-access-wf4xd\") pod \"observability-operator-d8bb48f5d-v7c8l\" (UID: \"714c939d-43ee-4564-851d-198cc78fbf7f\") " pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.453018 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm5vk\" (UniqueName: \"kubernetes.io/projected/8c2118f9-ee5c-4fb7-a734-f9fbc098c53b-kube-api-access-nm5vk\") pod \"obo-prometheus-operator-668cf9dfbb-k9rt9\" (UID: \"8c2118f9-ee5c-4fb7-a734-f9fbc098c53b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.453090 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84bb503d-0098-4342-bad3-6dc8e2ae325b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-4vxch\" (UID: \"84bb503d-0098-4342-bad3-6dc8e2ae325b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.453151 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/714c939d-43ee-4564-851d-198cc78fbf7f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-v7c8l\" (UID: \"714c939d-43ee-4564-851d-198cc78fbf7f\") " pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.453174 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cc522201-89c0-478e-8e63-4962e162b4e6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-c54zb\" (UID: \"cc522201-89c0-478e-8e63-4962e162b4e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.453200 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84bb503d-0098-4342-bad3-6dc8e2ae325b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-4vxch\" (UID: \"84bb503d-0098-4342-bad3-6dc8e2ae325b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.453237 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cc522201-89c0-478e-8e63-4962e162b4e6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-c54zb\" (UID: \"cc522201-89c0-478e-8e63-4962e162b4e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.466000 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cc522201-89c0-478e-8e63-4962e162b4e6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-c54zb\" (UID: \"cc522201-89c0-478e-8e63-4962e162b4e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.466043 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cc522201-89c0-478e-8e63-4962e162b4e6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-c54zb\" (UID: \"cc522201-89c0-478e-8e63-4962e162b4e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.466643 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-clc56"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.467973 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.472009 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-9ttw8" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.472203 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/84bb503d-0098-4342-bad3-6dc8e2ae325b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-4vxch\" (UID: \"84bb503d-0098-4342-bad3-6dc8e2ae325b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.476193 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-clc56"] Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.476802 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/84bb503d-0098-4342-bad3-6dc8e2ae325b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-644c864545-4vxch\" (UID: \"84bb503d-0098-4342-bad3-6dc8e2ae325b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.482599 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm5vk\" (UniqueName: \"kubernetes.io/projected/8c2118f9-ee5c-4fb7-a734-f9fbc098c53b-kube-api-access-nm5vk\") pod \"obo-prometheus-operator-668cf9dfbb-k9rt9\" (UID: \"8c2118f9-ee5c-4fb7-a734-f9fbc098c53b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.531317 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.548079 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.554220 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf4xd\" (UniqueName: \"kubernetes.io/projected/714c939d-43ee-4564-851d-198cc78fbf7f-kube-api-access-wf4xd\") pod \"observability-operator-d8bb48f5d-v7c8l\" (UID: \"714c939d-43ee-4564-851d-198cc78fbf7f\") " pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.554544 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdzf5\" (UniqueName: \"kubernetes.io/projected/c5c1bf2d-26d2-4df3-8c0a-fe99d9715603-kube-api-access-gdzf5\") pod \"perses-operator-5446b9c989-clc56\" (UID: \"c5c1bf2d-26d2-4df3-8c0a-fe99d9715603\") " pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.554661 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5c1bf2d-26d2-4df3-8c0a-fe99d9715603-openshift-service-ca\") pod \"perses-operator-5446b9c989-clc56\" (UID: \"c5c1bf2d-26d2-4df3-8c0a-fe99d9715603\") " pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.554714 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/714c939d-43ee-4564-851d-198cc78fbf7f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-v7c8l\" (UID: \"714c939d-43ee-4564-851d-198cc78fbf7f\") " pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.559468 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/714c939d-43ee-4564-851d-198cc78fbf7f-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-v7c8l\" (UID: \"714c939d-43ee-4564-851d-198cc78fbf7f\") " pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.573985 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf4xd\" (UniqueName: \"kubernetes.io/projected/714c939d-43ee-4564-851d-198cc78fbf7f-kube-api-access-wf4xd\") pod \"observability-operator-d8bb48f5d-v7c8l\" (UID: \"714c939d-43ee-4564-851d-198cc78fbf7f\") " pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.581984 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.656626 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5c1bf2d-26d2-4df3-8c0a-fe99d9715603-openshift-service-ca\") pod \"perses-operator-5446b9c989-clc56\" (UID: \"c5c1bf2d-26d2-4df3-8c0a-fe99d9715603\") " pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.656789 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdzf5\" (UniqueName: \"kubernetes.io/projected/c5c1bf2d-26d2-4df3-8c0a-fe99d9715603-kube-api-access-gdzf5\") pod \"perses-operator-5446b9c989-clc56\" (UID: \"c5c1bf2d-26d2-4df3-8c0a-fe99d9715603\") " pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.657703 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5c1bf2d-26d2-4df3-8c0a-fe99d9715603-openshift-service-ca\") pod \"perses-operator-5446b9c989-clc56\" (UID: \"c5c1bf2d-26d2-4df3-8c0a-fe99d9715603\") " pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.693591 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdzf5\" (UniqueName: \"kubernetes.io/projected/c5c1bf2d-26d2-4df3-8c0a-fe99d9715603-kube-api-access-gdzf5\") pod \"perses-operator-5446b9c989-clc56\" (UID: \"c5c1bf2d-26d2-4df3-8c0a-fe99d9715603\") " pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.698546 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:40 crc kubenswrapper[4876]: I1205 09:09:40.760947 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:41 crc kubenswrapper[4876]: I1205 09:09:41.597124 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9"] Dec 05 09:09:41 crc kubenswrapper[4876]: W1205 09:09:41.614743 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c2118f9_ee5c_4fb7_a734_f9fbc098c53b.slice/crio-bc450730472a2ee895b39526cfce6c212cc3bf22a29e105dd55e984824989bbe WatchSource:0}: Error finding container bc450730472a2ee895b39526cfce6c212cc3bf22a29e105dd55e984824989bbe: Status 404 returned error can't find the container with id bc450730472a2ee895b39526cfce6c212cc3bf22a29e105dd55e984824989bbe Dec 05 09:09:41 crc kubenswrapper[4876]: I1205 09:09:41.690290 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch"] Dec 05 09:09:41 crc kubenswrapper[4876]: W1205 09:09:41.868002 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc522201_89c0_478e_8e63_4962e162b4e6.slice/crio-31a7d7e1d8aee1fbcf901553d01e6d4eee4fe0ea0540be45fda8da4936de7b0c WatchSource:0}: Error finding container 31a7d7e1d8aee1fbcf901553d01e6d4eee4fe0ea0540be45fda8da4936de7b0c: Status 404 returned error can't find the container with id 31a7d7e1d8aee1fbcf901553d01e6d4eee4fe0ea0540be45fda8da4936de7b0c Dec 05 09:09:41 crc kubenswrapper[4876]: W1205 09:09:41.870862 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5c1bf2d_26d2_4df3_8c0a_fe99d9715603.slice/crio-e61fb183a6be89fbe42291caba0c96070ae7363cbcb6e900351b2599594460cc WatchSource:0}: Error finding container e61fb183a6be89fbe42291caba0c96070ae7363cbcb6e900351b2599594460cc: Status 404 returned error can't find the container with id e61fb183a6be89fbe42291caba0c96070ae7363cbcb6e900351b2599594460cc Dec 05 09:09:41 crc kubenswrapper[4876]: I1205 09:09:41.874834 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb"] Dec 05 09:09:41 crc kubenswrapper[4876]: W1205 09:09:41.877383 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod714c939d_43ee_4564_851d_198cc78fbf7f.slice/crio-2652686c24b0a06bf906d191371a1dfa10929970cb462631d91d57bf04475578 WatchSource:0}: Error finding container 2652686c24b0a06bf906d191371a1dfa10929970cb462631d91d57bf04475578: Status 404 returned error can't find the container with id 2652686c24b0a06bf906d191371a1dfa10929970cb462631d91d57bf04475578 Dec 05 09:09:41 crc kubenswrapper[4876]: I1205 09:09:41.885208 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-clc56"] Dec 05 09:09:41 crc kubenswrapper[4876]: I1205 09:09:41.895417 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-v7c8l"] Dec 05 09:09:42 crc kubenswrapper[4876]: I1205 09:09:42.300365 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-clc56" event={"ID":"c5c1bf2d-26d2-4df3-8c0a-fe99d9715603","Type":"ContainerStarted","Data":"e61fb183a6be89fbe42291caba0c96070ae7363cbcb6e900351b2599594460cc"} Dec 05 09:09:42 crc kubenswrapper[4876]: I1205 09:09:42.306856 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" event={"ID":"cc522201-89c0-478e-8e63-4962e162b4e6","Type":"ContainerStarted","Data":"31a7d7e1d8aee1fbcf901553d01e6d4eee4fe0ea0540be45fda8da4936de7b0c"} Dec 05 09:09:42 crc kubenswrapper[4876]: I1205 09:09:42.320938 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" event={"ID":"714c939d-43ee-4564-851d-198cc78fbf7f","Type":"ContainerStarted","Data":"2652686c24b0a06bf906d191371a1dfa10929970cb462631d91d57bf04475578"} Dec 05 09:09:42 crc kubenswrapper[4876]: I1205 09:09:42.326758 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" event={"ID":"8c2118f9-ee5c-4fb7-a734-f9fbc098c53b","Type":"ContainerStarted","Data":"bc450730472a2ee895b39526cfce6c212cc3bf22a29e105dd55e984824989bbe"} Dec 05 09:09:42 crc kubenswrapper[4876]: I1205 09:09:42.328363 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" event={"ID":"84bb503d-0098-4342-bad3-6dc8e2ae325b","Type":"ContainerStarted","Data":"a91ef7dd5717a7766e1bb21f961ac519ecc6418f04acf3af52784cbe37806fdf"} Dec 05 09:09:57 crc kubenswrapper[4876]: E1205 09:09:57.299090 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 05 09:09:57 crc kubenswrapper[4876]: E1205 09:09:57.299750 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-644c864545-4vxch_openshift-operators(84bb503d-0098-4342-bad3-6dc8e2ae325b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:09:57 crc kubenswrapper[4876]: E1205 09:09:57.300958 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" podUID="84bb503d-0098-4342-bad3-6dc8e2ae325b" Dec 05 09:09:57 crc kubenswrapper[4876]: E1205 09:09:57.341232 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 05 09:09:57 crc kubenswrapper[4876]: E1205 09:09:57.341404 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-644c864545-c54zb_openshift-operators(cc522201-89c0-478e-8e63-4962e162b4e6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:09:57 crc kubenswrapper[4876]: E1205 09:09:57.345189 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" podUID="cc522201-89c0-478e-8e63-4962e162b4e6" Dec 05 09:09:57 crc kubenswrapper[4876]: E1205 09:09:57.574325 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" podUID="cc522201-89c0-478e-8e63-4962e162b4e6" Dec 05 09:09:57 crc kubenswrapper[4876]: E1205 09:09:57.575037 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" podUID="84bb503d-0098-4342-bad3-6dc8e2ae325b" Dec 05 09:09:58 crc kubenswrapper[4876]: E1205 09:09:58.277136 4876 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3" Dec 05 09:09:58 crc kubenswrapper[4876]: E1205 09:09:58.277865 4876 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3,Command:[],Args:[--prometheus-config-reloader=$(RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER) --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOGC,Value:30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER,Value:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:1133c973c7472c665f910a722e19c8e2e27accb34b90fab67f14548627ce9c62,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nm5vk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-668cf9dfbb-k9rt9_openshift-operators(8c2118f9-ee5c-4fb7-a734-f9fbc098c53b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:09:58 crc kubenswrapper[4876]: E1205 09:09:58.279110 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" podUID="8c2118f9-ee5c-4fb7-a734-f9fbc098c53b" Dec 05 09:09:58 crc kubenswrapper[4876]: I1205 09:09:58.585845 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-clc56" event={"ID":"c5c1bf2d-26d2-4df3-8c0a-fe99d9715603","Type":"ContainerStarted","Data":"8ca5fcb1141847556caef803fb269d241a36a33eccf5fdc9cad972cda2ae86ac"} Dec 05 09:09:58 crc kubenswrapper[4876]: I1205 09:09:58.587803 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:09:58 crc kubenswrapper[4876]: I1205 09:09:58.588468 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" event={"ID":"714c939d-43ee-4564-851d-198cc78fbf7f","Type":"ContainerStarted","Data":"a0a065844701276780961fe61df9dd94b1c0c8f39400c26c8d43a1382424dab1"} Dec 05 09:09:58 crc kubenswrapper[4876]: I1205 09:09:58.588838 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:09:58 crc kubenswrapper[4876]: I1205 09:09:58.592123 4876 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-v7c8l container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.2:8081/healthz\": dial tcp 10.217.1.2:8081: connect: connection refused" start-of-body= Dec 05 09:09:58 crc kubenswrapper[4876]: I1205 09:09:58.592167 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" podUID="714c939d-43ee-4564-851d-198cc78fbf7f" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.2:8081/healthz\": dial tcp 10.217.1.2:8081: connect: connection refused" Dec 05 09:09:58 crc kubenswrapper[4876]: E1205 09:09:58.592495 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:203cf5b9dc1460f09e75f58d8b5cf7df5e57c18c8c6a41c14b5e8977d83263f3\\\"\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" podUID="8c2118f9-ee5c-4fb7-a734-f9fbc098c53b" Dec 05 09:09:58 crc kubenswrapper[4876]: I1205 09:09:58.632115 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" podStartSLOduration=2.210335304 podStartE2EDuration="18.632096491s" podCreationTimestamp="2025-12-05 09:09:40 +0000 UTC" firstStartedPulling="2025-12-05 09:09:41.881473231 +0000 UTC m=+3246.370137863" lastFinishedPulling="2025-12-05 09:09:58.303234428 +0000 UTC m=+3262.791899050" observedRunningTime="2025-12-05 09:09:58.630327223 +0000 UTC m=+3263.118991845" watchObservedRunningTime="2025-12-05 09:09:58.632096491 +0000 UTC m=+3263.120761113" Dec 05 09:09:58 crc kubenswrapper[4876]: I1205 09:09:58.636445 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-clc56" podStartSLOduration=2.229847623 podStartE2EDuration="18.636426408s" podCreationTimestamp="2025-12-05 09:09:40 +0000 UTC" firstStartedPulling="2025-12-05 09:09:41.872985261 +0000 UTC m=+3246.361649883" lastFinishedPulling="2025-12-05 09:09:58.279564046 +0000 UTC m=+3262.768228668" observedRunningTime="2025-12-05 09:09:58.611355928 +0000 UTC m=+3263.100020570" watchObservedRunningTime="2025-12-05 09:09:58.636426408 +0000 UTC m=+3263.125091030" Dec 05 09:09:59 crc kubenswrapper[4876]: I1205 09:09:59.596356 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-v7c8l" Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.120567 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.124022 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-api" containerID="cri-o://17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872" gracePeriod=30 Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.124064 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-listener" containerID="cri-o://7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15" gracePeriod=30 Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.124179 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-notifier" containerID="cri-o://445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c" gracePeriod=30 Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.124240 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-evaluator" containerID="cri-o://01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e" gracePeriod=30 Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.645296 4876 generic.go:334] "Generic (PLEG): container finished" podID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerID="01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e" exitCode=0 Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.645574 4876 generic.go:334] "Generic (PLEG): container finished" podID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerID="17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872" exitCode=0 Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.645385 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerDied","Data":"01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e"} Dec 05 09:10:04 crc kubenswrapper[4876]: I1205 09:10:04.645614 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerDied","Data":"17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872"} Dec 05 09:10:10 crc kubenswrapper[4876]: I1205 09:10:10.694837 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" event={"ID":"84bb503d-0098-4342-bad3-6dc8e2ae325b","Type":"ContainerStarted","Data":"289ed6187ea12a994c591cebb6c6dc7d00c8c5d23bee5d006ec2e6bbbcac3b6b"} Dec 05 09:10:10 crc kubenswrapper[4876]: I1205 09:10:10.723027 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-4vxch" podStartSLOduration=2.456499732 podStartE2EDuration="30.7230065s" podCreationTimestamp="2025-12-05 09:09:40 +0000 UTC" firstStartedPulling="2025-12-05 09:09:41.696055211 +0000 UTC m=+3246.184719833" lastFinishedPulling="2025-12-05 09:10:09.962561979 +0000 UTC m=+3274.451226601" observedRunningTime="2025-12-05 09:10:10.713776 +0000 UTC m=+3275.202440662" watchObservedRunningTime="2025-12-05 09:10:10.7230065 +0000 UTC m=+3275.211671122" Dec 05 09:10:10 crc kubenswrapper[4876]: I1205 09:10:10.771576 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-clc56" Dec 05 09:10:11 crc kubenswrapper[4876]: I1205 09:10:11.704656 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" event={"ID":"cc522201-89c0-478e-8e63-4962e162b4e6","Type":"ContainerStarted","Data":"6c500f999158c0711743615aa4a0b7a567c1b48c94b8b1e5f12790c6bd550362"} Dec 05 09:10:11 crc kubenswrapper[4876]: I1205 09:10:11.706509 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" event={"ID":"8c2118f9-ee5c-4fb7-a734-f9fbc098c53b","Type":"ContainerStarted","Data":"e4b1b62c9dc6d488c7dc28875d61b3ce33cf65ae1fac63289417e58e01ce0e48"} Dec 05 09:10:11 crc kubenswrapper[4876]: I1205 09:10:11.726577 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-644c864545-c54zb" podStartSLOduration=-9223372005.128223 podStartE2EDuration="31.726552829s" podCreationTimestamp="2025-12-05 09:09:40 +0000 UTC" firstStartedPulling="2025-12-05 09:09:41.872494458 +0000 UTC m=+3246.361159080" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:10:11.721138942 +0000 UTC m=+3276.209803564" watchObservedRunningTime="2025-12-05 09:10:11.726552829 +0000 UTC m=+3276.215217451" Dec 05 09:10:11 crc kubenswrapper[4876]: I1205 09:10:11.762084 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-k9rt9" podStartSLOduration=2.427574229 podStartE2EDuration="31.762058962s" podCreationTimestamp="2025-12-05 09:09:40 +0000 UTC" firstStartedPulling="2025-12-05 09:09:41.631357736 +0000 UTC m=+3246.120022358" lastFinishedPulling="2025-12-05 09:10:10.965842469 +0000 UTC m=+3275.454507091" observedRunningTime="2025-12-05 09:10:11.757023245 +0000 UTC m=+3276.245687877" watchObservedRunningTime="2025-12-05 09:10:11.762058962 +0000 UTC m=+3276.250723614" Dec 05 09:10:13 crc kubenswrapper[4876]: I1205 09:10:13.727814 4876 generic.go:334] "Generic (PLEG): container finished" podID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerID="445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c" exitCode=0 Dec 05 09:10:13 crc kubenswrapper[4876]: I1205 09:10:13.727950 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerDied","Data":"445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c"} Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.562988 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.647226 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-config-data\") pod \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.647325 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-combined-ca-bundle\") pod \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.647369 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-scripts\") pod \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.647400 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfmxl\" (UniqueName: \"kubernetes.io/projected/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-kube-api-access-rfmxl\") pod \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.647432 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-public-tls-certs\") pod \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.647643 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-internal-tls-certs\") pod \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\" (UID: \"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e\") " Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.662670 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-scripts" (OuterVolumeSpecName: "scripts") pod "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" (UID: "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.680807 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-kube-api-access-rfmxl" (OuterVolumeSpecName: "kube-api-access-rfmxl") pod "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" (UID: "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e"). InnerVolumeSpecName "kube-api-access-rfmxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.752212 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.752249 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfmxl\" (UniqueName: \"kubernetes.io/projected/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-kube-api-access-rfmxl\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.761573 4876 generic.go:334] "Generic (PLEG): container finished" podID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerID="7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15" exitCode=0 Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.761649 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerDied","Data":"7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15"} Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.761682 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"0985a8e0-0ae3-42ad-b0a3-3011c24bb20e","Type":"ContainerDied","Data":"2437d7ccb64300e95892f4fa2fbe5372a072e9a66bf962385fc428da8573a8a2"} Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.761700 4876 scope.go:117] "RemoveContainer" containerID="7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.762050 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.822837 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" (UID: "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.855681 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.872060 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" (UID: "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.876102 4876 scope.go:117] "RemoveContainer" containerID="445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.910221 4876 scope.go:117] "RemoveContainer" containerID="01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.921241 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-config-data" (OuterVolumeSpecName: "config-data") pod "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" (UID: "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.942220 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" (UID: "0985a8e0-0ae3-42ad-b0a3-3011c24bb20e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.958237 4876 scope.go:117] "RemoveContainer" containerID="17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.960101 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.960135 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:14 crc kubenswrapper[4876]: I1205 09:10:14.960147 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.013505 4876 scope.go:117] "RemoveContainer" containerID="7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15" Dec 05 09:10:15 crc kubenswrapper[4876]: E1205 09:10:15.014226 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15\": container with ID starting with 7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15 not found: ID does not exist" containerID="7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.014260 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15"} err="failed to get container status \"7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15\": rpc error: code = NotFound desc = could not find container \"7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15\": container with ID starting with 7d9eb12384ae7d951ca4f46de46053285725d4ef7dcf46cbb0cd0517c9e31d15 not found: ID does not exist" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.014299 4876 scope.go:117] "RemoveContainer" containerID="445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c" Dec 05 09:10:15 crc kubenswrapper[4876]: E1205 09:10:15.018249 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c\": container with ID starting with 445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c not found: ID does not exist" containerID="445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.018293 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c"} err="failed to get container status \"445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c\": rpc error: code = NotFound desc = could not find container \"445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c\": container with ID starting with 445bebc92cb8c4155ff6e1e5bcf9109784be4ab81b2c80a324ffc48f9ba58b3c not found: ID does not exist" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.018323 4876 scope.go:117] "RemoveContainer" containerID="01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e" Dec 05 09:10:15 crc kubenswrapper[4876]: E1205 09:10:15.018705 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e\": container with ID starting with 01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e not found: ID does not exist" containerID="01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.018740 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e"} err="failed to get container status \"01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e\": rpc error: code = NotFound desc = could not find container \"01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e\": container with ID starting with 01d4e9dd5ef428d3a89a2e8477c906eb24efe63952fc2126d7742b4fe526c04e not found: ID does not exist" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.018760 4876 scope.go:117] "RemoveContainer" containerID="17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872" Dec 05 09:10:15 crc kubenswrapper[4876]: E1205 09:10:15.019070 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872\": container with ID starting with 17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872 not found: ID does not exist" containerID="17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.019097 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872"} err="failed to get container status \"17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872\": rpc error: code = NotFound desc = could not find container \"17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872\": container with ID starting with 17e22dd6f123177c715a35622fab9210b1364ab2f9a4ccdac4b2100e63956872 not found: ID does not exist" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.121093 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.142438 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.199884 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 05 09:10:15 crc kubenswrapper[4876]: E1205 09:10:15.200401 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-evaluator" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.200420 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-evaluator" Dec 05 09:10:15 crc kubenswrapper[4876]: E1205 09:10:15.200449 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-listener" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.200457 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-listener" Dec 05 09:10:15 crc kubenswrapper[4876]: E1205 09:10:15.200472 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-api" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.200477 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-api" Dec 05 09:10:15 crc kubenswrapper[4876]: E1205 09:10:15.200489 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-notifier" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.200494 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-notifier" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.200670 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-notifier" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.200685 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-api" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.200694 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-listener" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.200708 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" containerName="aodh-evaluator" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.202595 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.206235 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.206263 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tqf85" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.206273 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.206356 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.206809 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.212399 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.265773 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-config-data\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.265988 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-public-tls-certs\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.266019 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.266106 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5mkc\" (UniqueName: \"kubernetes.io/projected/4601390b-7a61-460d-8f54-83886291c0c2-kube-api-access-q5mkc\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.266232 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-internal-tls-certs\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.266310 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-scripts\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.370712 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-public-tls-certs\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.371340 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.371411 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5mkc\" (UniqueName: \"kubernetes.io/projected/4601390b-7a61-460d-8f54-83886291c0c2-kube-api-access-q5mkc\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.371460 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-internal-tls-certs\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.371487 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-scripts\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.371581 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-config-data\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.376689 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-public-tls-certs\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.381679 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.382284 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-internal-tls-certs\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.383477 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-config-data\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.386311 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-scripts\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.407649 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5mkc\" (UniqueName: \"kubernetes.io/projected/4601390b-7a61-460d-8f54-83886291c0c2-kube-api-access-q5mkc\") pod \"aodh-0\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.518648 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.540783 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8jhp7"] Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.543574 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.564116 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.566060 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.576399 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.576659 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-j8pn7" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.576771 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.576775 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.576875 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.626922 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8jhp7"] Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681197 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2jjq\" (UniqueName: \"kubernetes.io/projected/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-kube-api-access-c2jjq\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681274 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681312 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681336 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-catalog-content\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681394 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a9faf08a-6b0e-435c-9810-d94212aa82ad-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681451 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbfm2\" (UniqueName: \"kubernetes.io/projected/a9faf08a-6b0e-435c-9810-d94212aa82ad-kube-api-access-vbfm2\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681525 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681585 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a9faf08a-6b0e-435c-9810-d94212aa82ad-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681606 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/a9faf08a-6b0e-435c-9810-d94212aa82ad-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.681629 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-utilities\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.717094 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.783426 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:15 crc kubenswrapper[4876]: I1205 09:10:15.783485 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.783518 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-catalog-content\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.783601 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a9faf08a-6b0e-435c-9810-d94212aa82ad-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.783638 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbfm2\" (UniqueName: \"kubernetes.io/projected/a9faf08a-6b0e-435c-9810-d94212aa82ad-kube-api-access-vbfm2\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.783664 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.783682 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a9faf08a-6b0e-435c-9810-d94212aa82ad-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.783725 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/a9faf08a-6b0e-435c-9810-d94212aa82ad-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.783745 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-utilities\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.783795 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2jjq\" (UniqueName: \"kubernetes.io/projected/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-kube-api-access-c2jjq\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.792739 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-catalog-content\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.796115 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a9faf08a-6b0e-435c-9810-d94212aa82ad-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.796647 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/a9faf08a-6b0e-435c-9810-d94212aa82ad-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.796953 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-utilities\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.808444 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.815410 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.815743 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a9faf08a-6b0e-435c-9810-d94212aa82ad-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.816840 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a9faf08a-6b0e-435c-9810-d94212aa82ad-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.870928 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbfm2\" (UniqueName: \"kubernetes.io/projected/a9faf08a-6b0e-435c-9810-d94212aa82ad-kube-api-access-vbfm2\") pod \"alertmanager-metric-storage-0\" (UID: \"a9faf08a-6b0e-435c-9810-d94212aa82ad\") " pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:15.883755 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2jjq\" (UniqueName: \"kubernetes.io/projected/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-kube-api-access-c2jjq\") pod \"redhat-operators-8jhp7\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.132864 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0985a8e0-0ae3-42ad-b0a3-3011c24bb20e" path="/var/lib/kubelet/pods/0985a8e0-0ae3-42ad-b0a3-3011c24bb20e/volumes" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.162139 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.185849 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.597522 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.797006 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.799995 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.808810 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.809095 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.809265 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.809434 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.809674 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-zx6qt" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.813716 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.813975 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.826082 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerStarted","Data":"5fb15807ccc9d4467fcee9f436056fb4c4246d9ea870deabd1543613009c861c"} Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.844156 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.844211 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.844264 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kql6m\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-kube-api-access-kql6m\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.844293 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.844319 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.844359 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.844376 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.844400 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.868417 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8jhp7"] Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.946406 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kql6m\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-kube-api-access-kql6m\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.946483 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.946546 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.946590 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.946635 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.946660 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.946793 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.946826 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.949237 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.951132 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.956940 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.957361 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.957454 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.966947 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.968535 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.980985 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kql6m\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-kube-api-access-kql6m\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:16 crc kubenswrapper[4876]: I1205 09:10:16.996500 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.024335 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.135676 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.694344 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:10:17 crc kubenswrapper[4876]: W1205 09:10:17.708388 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cb6f8ab_8fcc_4518_88e7_4f479a828d00.slice/crio-68b13db6025efff889a843e12fbda3344f880ccd77527af31cae918d9a4fa9b2 WatchSource:0}: Error finding container 68b13db6025efff889a843e12fbda3344f880ccd77527af31cae918d9a4fa9b2: Status 404 returned error can't find the container with id 68b13db6025efff889a843e12fbda3344f880ccd77527af31cae918d9a4fa9b2 Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.835290 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerStarted","Data":"68b13db6025efff889a843e12fbda3344f880ccd77527af31cae918d9a4fa9b2"} Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.836711 4876 generic.go:334] "Generic (PLEG): container finished" podID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerID="3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062" exitCode=0 Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.836750 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jhp7" event={"ID":"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d","Type":"ContainerDied","Data":"3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062"} Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.836766 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jhp7" event={"ID":"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d","Type":"ContainerStarted","Data":"f46630c989aa3a5775169c1712bcaf4af0d00e428d565423881f56d1ed021f9d"} Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.840272 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerStarted","Data":"4b5da10bf892b403d3e86d53172428c255a8471861df08374e19429d62db2128"} Dec 05 09:10:17 crc kubenswrapper[4876]: I1205 09:10:17.842624 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a9faf08a-6b0e-435c-9810-d94212aa82ad","Type":"ContainerStarted","Data":"63006c642a38892a553e8bbf3d4a23ac3f06f976bd62b35e73f6aa877a596487"} Dec 05 09:10:18 crc kubenswrapper[4876]: I1205 09:10:18.861599 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerStarted","Data":"29f2e324112ee15af05481b82cd17d1f42324587a004b8d2f8b5d497d525b462"} Dec 05 09:10:18 crc kubenswrapper[4876]: I1205 09:10:18.868825 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jhp7" event={"ID":"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d","Type":"ContainerStarted","Data":"29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586"} Dec 05 09:10:20 crc kubenswrapper[4876]: I1205 09:10:20.892593 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerStarted","Data":"4aafdc439d48b94b197c3f1b87d0c22892917fbd0a9fb609fc0d84730b8c99db"} Dec 05 09:10:23 crc kubenswrapper[4876]: I1205 09:10:23.923714 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerStarted","Data":"08f3d78e7146f8ec22955491551da094db8aee633e3449dcd3a46682ebbc1ebc"} Dec 05 09:10:23 crc kubenswrapper[4876]: I1205 09:10:23.925766 4876 generic.go:334] "Generic (PLEG): container finished" podID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerID="29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586" exitCode=0 Dec 05 09:10:23 crc kubenswrapper[4876]: I1205 09:10:23.925820 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jhp7" event={"ID":"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d","Type":"ContainerDied","Data":"29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586"} Dec 05 09:10:23 crc kubenswrapper[4876]: I1205 09:10:23.931456 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerStarted","Data":"95c5aff4c7ebdbb3bfc46076190b6cbd4739b6d84e3f7a3dfb0022e996561207"} Dec 05 09:10:23 crc kubenswrapper[4876]: I1205 09:10:23.933390 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a9faf08a-6b0e-435c-9810-d94212aa82ad","Type":"ContainerStarted","Data":"9e343488c744351d06fdcad00f099e4a36d74f48f87bc27ae252861728b8a7d6"} Dec 05 09:10:23 crc kubenswrapper[4876]: I1205 09:10:23.999024 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.882130913 podStartE2EDuration="8.999008656s" podCreationTimestamp="2025-12-05 09:10:15 +0000 UTC" firstStartedPulling="2025-12-05 09:10:16.606854402 +0000 UTC m=+3281.095519024" lastFinishedPulling="2025-12-05 09:10:22.723732145 +0000 UTC m=+3287.212396767" observedRunningTime="2025-12-05 09:10:23.99805272 +0000 UTC m=+3288.486717342" watchObservedRunningTime="2025-12-05 09:10:23.999008656 +0000 UTC m=+3288.487673288" Dec 05 09:10:24 crc kubenswrapper[4876]: I1205 09:10:24.946048 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jhp7" event={"ID":"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d","Type":"ContainerStarted","Data":"9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79"} Dec 05 09:10:24 crc kubenswrapper[4876]: I1205 09:10:24.992743 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8jhp7" podStartSLOduration=3.306930507 podStartE2EDuration="9.992716777s" podCreationTimestamp="2025-12-05 09:10:15 +0000 UTC" firstStartedPulling="2025-12-05 09:10:17.838329063 +0000 UTC m=+3282.326993685" lastFinishedPulling="2025-12-05 09:10:24.524115323 +0000 UTC m=+3289.012779955" observedRunningTime="2025-12-05 09:10:24.985032188 +0000 UTC m=+3289.473696810" watchObservedRunningTime="2025-12-05 09:10:24.992716777 +0000 UTC m=+3289.481381409" Dec 05 09:10:26 crc kubenswrapper[4876]: I1205 09:10:26.162307 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:26 crc kubenswrapper[4876]: I1205 09:10:26.162639 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:27 crc kubenswrapper[4876]: I1205 09:10:27.213455 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8jhp7" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="registry-server" probeResult="failure" output=< Dec 05 09:10:27 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 09:10:27 crc kubenswrapper[4876]: > Dec 05 09:10:30 crc kubenswrapper[4876]: E1205 09:10:30.395282 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cb6f8ab_8fcc_4518_88e7_4f479a828d00.slice/crio-08f3d78e7146f8ec22955491551da094db8aee633e3449dcd3a46682ebbc1ebc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cb6f8ab_8fcc_4518_88e7_4f479a828d00.slice/crio-conmon-08f3d78e7146f8ec22955491551da094db8aee633e3449dcd3a46682ebbc1ebc.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:10:31 crc kubenswrapper[4876]: I1205 09:10:31.000310 4876 generic.go:334] "Generic (PLEG): container finished" podID="a9faf08a-6b0e-435c-9810-d94212aa82ad" containerID="9e343488c744351d06fdcad00f099e4a36d74f48f87bc27ae252861728b8a7d6" exitCode=0 Dec 05 09:10:31 crc kubenswrapper[4876]: I1205 09:10:31.000405 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a9faf08a-6b0e-435c-9810-d94212aa82ad","Type":"ContainerDied","Data":"9e343488c744351d06fdcad00f099e4a36d74f48f87bc27ae252861728b8a7d6"} Dec 05 09:10:31 crc kubenswrapper[4876]: I1205 09:10:31.003135 4876 generic.go:334] "Generic (PLEG): container finished" podID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerID="08f3d78e7146f8ec22955491551da094db8aee633e3449dcd3a46682ebbc1ebc" exitCode=0 Dec 05 09:10:31 crc kubenswrapper[4876]: I1205 09:10:31.003179 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerDied","Data":"08f3d78e7146f8ec22955491551da094db8aee633e3449dcd3a46682ebbc1ebc"} Dec 05 09:10:34 crc kubenswrapper[4876]: I1205 09:10:34.045435 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a9faf08a-6b0e-435c-9810-d94212aa82ad","Type":"ContainerStarted","Data":"e1d2439973696df016ea13bd72d98050e9bbde4d95741e583305e2553882ea83"} Dec 05 09:10:37 crc kubenswrapper[4876]: I1205 09:10:37.222331 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8jhp7" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="registry-server" probeResult="failure" output=< Dec 05 09:10:37 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 09:10:37 crc kubenswrapper[4876]: > Dec 05 09:10:38 crc kubenswrapper[4876]: I1205 09:10:38.109629 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerStarted","Data":"9e3064522f5bf0b86fc352814475b6b5bd89ce900d8725567559ba6497ea074f"} Dec 05 09:10:42 crc kubenswrapper[4876]: I1205 09:10:42.144691 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"a9faf08a-6b0e-435c-9810-d94212aa82ad","Type":"ContainerStarted","Data":"7df0c636389dd734d93f86d79e417797bcea4143d14ea89a8a9c084e5b684f9f"} Dec 05 09:10:42 crc kubenswrapper[4876]: I1205 09:10:42.145626 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:42 crc kubenswrapper[4876]: I1205 09:10:42.149328 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 05 09:10:42 crc kubenswrapper[4876]: I1205 09:10:42.177212 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=10.706497433 podStartE2EDuration="27.177185956s" podCreationTimestamp="2025-12-05 09:10:15 +0000 UTC" firstStartedPulling="2025-12-05 09:10:17.020434193 +0000 UTC m=+3281.509098815" lastFinishedPulling="2025-12-05 09:10:33.491122706 +0000 UTC m=+3297.979787338" observedRunningTime="2025-12-05 09:10:42.171288626 +0000 UTC m=+3306.659953298" watchObservedRunningTime="2025-12-05 09:10:42.177185956 +0000 UTC m=+3306.665850598" Dec 05 09:10:46 crc kubenswrapper[4876]: I1205 09:10:46.213546 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:46 crc kubenswrapper[4876]: I1205 09:10:46.274702 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:46 crc kubenswrapper[4876]: I1205 09:10:46.724260 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8jhp7"] Dec 05 09:10:48 crc kubenswrapper[4876]: I1205 09:10:48.204400 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerStarted","Data":"ca4ddcc6807b9eb6d5fa6e5b73997f5428e0f1b481da964ed040ee650c089911"} Dec 05 09:10:48 crc kubenswrapper[4876]: I1205 09:10:48.204554 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8jhp7" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="registry-server" containerID="cri-o://9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79" gracePeriod=2 Dec 05 09:10:48 crc kubenswrapper[4876]: I1205 09:10:48.746267 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:48 crc kubenswrapper[4876]: I1205 09:10:48.942807 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2jjq\" (UniqueName: \"kubernetes.io/projected/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-kube-api-access-c2jjq\") pod \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " Dec 05 09:10:48 crc kubenswrapper[4876]: I1205 09:10:48.942862 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-catalog-content\") pod \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " Dec 05 09:10:48 crc kubenswrapper[4876]: I1205 09:10:48.942908 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-utilities\") pod \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\" (UID: \"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d\") " Dec 05 09:10:48 crc kubenswrapper[4876]: I1205 09:10:48.944187 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-utilities" (OuterVolumeSpecName: "utilities") pod "8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" (UID: "8b7f379e-310b-4a49-bd9e-b6e5d8193a9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:10:48 crc kubenswrapper[4876]: I1205 09:10:48.950324 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-kube-api-access-c2jjq" (OuterVolumeSpecName: "kube-api-access-c2jjq") pod "8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" (UID: "8b7f379e-310b-4a49-bd9e-b6e5d8193a9d"). InnerVolumeSpecName "kube-api-access-c2jjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.046058 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2jjq\" (UniqueName: \"kubernetes.io/projected/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-kube-api-access-c2jjq\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.046091 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.050730 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" (UID: "8b7f379e-310b-4a49-bd9e-b6e5d8193a9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.147380 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.216411 4876 generic.go:334] "Generic (PLEG): container finished" podID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerID="9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79" exitCode=0 Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.216462 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jhp7" event={"ID":"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d","Type":"ContainerDied","Data":"9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79"} Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.216501 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8jhp7" event={"ID":"8b7f379e-310b-4a49-bd9e-b6e5d8193a9d","Type":"ContainerDied","Data":"f46630c989aa3a5775169c1712bcaf4af0d00e428d565423881f56d1ed021f9d"} Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.216523 4876 scope.go:117] "RemoveContainer" containerID="9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.216528 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8jhp7" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.240953 4876 scope.go:117] "RemoveContainer" containerID="29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.254213 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8jhp7"] Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.264009 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8jhp7"] Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.283064 4876 scope.go:117] "RemoveContainer" containerID="3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.321505 4876 scope.go:117] "RemoveContainer" containerID="9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79" Dec 05 09:10:49 crc kubenswrapper[4876]: E1205 09:10:49.324966 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79\": container with ID starting with 9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79 not found: ID does not exist" containerID="9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.325018 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79"} err="failed to get container status \"9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79\": rpc error: code = NotFound desc = could not find container \"9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79\": container with ID starting with 9230002e3b5fee5e3f3e1c0ad398a32bd2860ee1b02d5bea5133add3293aec79 not found: ID does not exist" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.325051 4876 scope.go:117] "RemoveContainer" containerID="29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586" Dec 05 09:10:49 crc kubenswrapper[4876]: E1205 09:10:49.327039 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586\": container with ID starting with 29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586 not found: ID does not exist" containerID="29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.327112 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586"} err="failed to get container status \"29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586\": rpc error: code = NotFound desc = could not find container \"29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586\": container with ID starting with 29fb8fda1c248624446cc1c466f09758bef3e3eaeb1c5d43f683ba235ed45586 not found: ID does not exist" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.327248 4876 scope.go:117] "RemoveContainer" containerID="3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062" Dec 05 09:10:49 crc kubenswrapper[4876]: E1205 09:10:49.328020 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062\": container with ID starting with 3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062 not found: ID does not exist" containerID="3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.328055 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062"} err="failed to get container status \"3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062\": rpc error: code = NotFound desc = could not find container \"3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062\": container with ID starting with 3fdf2c0b038dd73d78c72924ee9602f08bc31224dc74a6da64febcc69bd79062 not found: ID does not exist" Dec 05 09:10:49 crc kubenswrapper[4876]: I1205 09:10:49.843152 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" path="/var/lib/kubelet/pods/8b7f379e-310b-4a49-bd9e-b6e5d8193a9d/volumes" Dec 05 09:10:51 crc kubenswrapper[4876]: I1205 09:10:51.244706 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerStarted","Data":"96c0a945ba122d80746127a1d30427b8322d9a1676dab585fcfd5627935ed4b3"} Dec 05 09:10:51 crc kubenswrapper[4876]: I1205 09:10:51.281662 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.009687142 podStartE2EDuration="36.281640988s" podCreationTimestamp="2025-12-05 09:10:15 +0000 UTC" firstStartedPulling="2025-12-05 09:10:17.712792837 +0000 UTC m=+3282.201457459" lastFinishedPulling="2025-12-05 09:10:50.984746683 +0000 UTC m=+3315.473411305" observedRunningTime="2025-12-05 09:10:51.274288919 +0000 UTC m=+3315.762953551" watchObservedRunningTime="2025-12-05 09:10:51.281640988 +0000 UTC m=+3315.770305610" Dec 05 09:10:52 crc kubenswrapper[4876]: I1205 09:10:52.136992 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:02 crc kubenswrapper[4876]: I1205 09:11:02.136329 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:02 crc kubenswrapper[4876]: I1205 09:11:02.139292 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:02 crc kubenswrapper[4876]: I1205 09:11:02.506619 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.303744 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.304366 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="bf04a5a6-de08-4983-8986-c0c949090943" containerName="openstackclient" containerID="cri-o://c63b6e6c01b11dcf2df1d4790bc24afdbae1f0cde40168fb4e95ee8fc277a5b0" gracePeriod=2 Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.313192 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.328195 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 09:11:04 crc kubenswrapper[4876]: E1205 09:11:04.328664 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf04a5a6-de08-4983-8986-c0c949090943" containerName="openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.328692 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf04a5a6-de08-4983-8986-c0c949090943" containerName="openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: E1205 09:11:04.328711 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="extract-utilities" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.328718 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="extract-utilities" Dec 05 09:11:04 crc kubenswrapper[4876]: E1205 09:11:04.328753 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="extract-content" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.328759 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="extract-content" Dec 05 09:11:04 crc kubenswrapper[4876]: E1205 09:11:04.328771 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="registry-server" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.328777 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="registry-server" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.328998 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf04a5a6-de08-4983-8986-c0c949090943" containerName="openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.329025 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b7f379e-310b-4a49-bd9e-b6e5d8193a9d" containerName="registry-server" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.329713 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.332423 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf04a5a6-de08-4983-8986-c0c949090943" podUID="ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.337067 4876 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:11:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:11:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:11:04Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:11:04Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\\\",\\\"name\\\":\\\"combined-ca-bundle\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4jhzx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:11:04Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.347429 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 05 09:11:04 crc kubenswrapper[4876]: E1205 09:11:04.352555 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-4jhzx openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.358528 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.361116 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jhzx\" (UniqueName: \"kubernetes.io/projected/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-kube-api-access-4jhzx\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.361242 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.361302 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.361327 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config-secret\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.384932 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.386273 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.392658 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" podUID="17ad89b3-21e5-4db2-8d47-29248aa9e054" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.453602 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.463107 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jhzx\" (UniqueName: \"kubernetes.io/projected/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-kube-api-access-4jhzx\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.463168 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/17ad89b3-21e5-4db2-8d47-29248aa9e054-openstack-config\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.463207 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/17ad89b3-21e5-4db2-8d47-29248aa9e054-openstack-config-secret\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.463296 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.463353 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.463376 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17ad89b3-21e5-4db2-8d47-29248aa9e054-combined-ca-bundle\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.463393 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config-secret\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.463411 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmvmw\" (UniqueName: \"kubernetes.io/projected/17ad89b3-21e5-4db2-8d47-29248aa9e054-kube-api-access-wmvmw\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.464565 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: E1205 09:11:04.468850 4876 projected.go:194] Error preparing data for projected volume kube-api-access-4jhzx for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f) does not match the UID in record. The object might have been deleted and then recreated Dec 05 09:11:04 crc kubenswrapper[4876]: E1205 09:11:04.468940 4876 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-kube-api-access-4jhzx podName:ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f nodeName:}" failed. No retries permitted until 2025-12-05 09:11:04.968918456 +0000 UTC m=+3329.457583078 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-4jhzx" (UniqueName: "kubernetes.io/projected/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-kube-api-access-4jhzx") pod "openstackclient" (UID: "ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f) does not match the UID in record. The object might have been deleted and then recreated Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.472269 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.484465 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config-secret\") pod \"openstackclient\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.523818 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.527104 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" podUID="17ad89b3-21e5-4db2-8d47-29248aa9e054" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.556995 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.560886 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" podUID="17ad89b3-21e5-4db2-8d47-29248aa9e054" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.565184 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17ad89b3-21e5-4db2-8d47-29248aa9e054-combined-ca-bundle\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.565220 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmvmw\" (UniqueName: \"kubernetes.io/projected/17ad89b3-21e5-4db2-8d47-29248aa9e054-kube-api-access-wmvmw\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.565292 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/17ad89b3-21e5-4db2-8d47-29248aa9e054-openstack-config\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.565324 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/17ad89b3-21e5-4db2-8d47-29248aa9e054-openstack-config-secret\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.566868 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/17ad89b3-21e5-4db2-8d47-29248aa9e054-openstack-config\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.568702 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/17ad89b3-21e5-4db2-8d47-29248aa9e054-openstack-config-secret\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.569290 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17ad89b3-21e5-4db2-8d47-29248aa9e054-combined-ca-bundle\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.582670 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmvmw\" (UniqueName: \"kubernetes.io/projected/17ad89b3-21e5-4db2-8d47-29248aa9e054-kube-api-access-wmvmw\") pod \"openstackclient\" (UID: \"17ad89b3-21e5-4db2-8d47-29248aa9e054\") " pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.611448 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.613759 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-listener" containerID="cri-o://95c5aff4c7ebdbb3bfc46076190b6cbd4739b6d84e3f7a3dfb0022e996561207" gracePeriod=30 Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.613976 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-notifier" containerID="cri-o://4aafdc439d48b94b197c3f1b87d0c22892917fbd0a9fb609fc0d84730b8c99db" gracePeriod=30 Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.614024 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-evaluator" containerID="cri-o://29f2e324112ee15af05481b82cd17d1f42324587a004b8d2f8b5d497d525b462" gracePeriod=30 Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.613812 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-api" containerID="cri-o://4b5da10bf892b403d3e86d53172428c255a8471861df08374e19429d62db2128" gracePeriod=30 Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.666626 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config-secret\") pod \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.666712 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-combined-ca-bundle\") pod \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.666889 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config\") pod \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\" (UID: \"ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f\") " Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.667514 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jhzx\" (UniqueName: \"kubernetes.io/projected/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-kube-api-access-4jhzx\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.667657 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" (UID: "ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.670882 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" (UID: "ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.672994 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" (UID: "ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.712976 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.769243 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.769585 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:04 crc kubenswrapper[4876]: I1205 09:11:04.769597 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.263978 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.496916 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.497663 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="prometheus" containerID="cri-o://9e3064522f5bf0b86fc352814475b6b5bd89ce900d8725567559ba6497ea074f" gracePeriod=600 Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.497825 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="thanos-sidecar" containerID="cri-o://96c0a945ba122d80746127a1d30427b8322d9a1676dab585fcfd5627935ed4b3" gracePeriod=600 Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.497876 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="config-reloader" containerID="cri-o://ca4ddcc6807b9eb6d5fa6e5b73997f5428e0f1b481da964ed040ee650c089911" gracePeriod=600 Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.537375 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"17ad89b3-21e5-4db2-8d47-29248aa9e054","Type":"ContainerStarted","Data":"69e8a0bfa84340272117bac39cd95f131e5f8e22e50e73a409c6226267b01060"} Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.537418 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"17ad89b3-21e5-4db2-8d47-29248aa9e054","Type":"ContainerStarted","Data":"d566f44525290803fa3814ef1595f54d5422458225078b61328097382cafaad3"} Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.541214 4876 generic.go:334] "Generic (PLEG): container finished" podID="4601390b-7a61-460d-8f54-83886291c0c2" containerID="29f2e324112ee15af05481b82cd17d1f42324587a004b8d2f8b5d497d525b462" exitCode=0 Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.541243 4876 generic.go:334] "Generic (PLEG): container finished" podID="4601390b-7a61-460d-8f54-83886291c0c2" containerID="4b5da10bf892b403d3e86d53172428c255a8471861df08374e19429d62db2128" exitCode=0 Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.541295 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.541290 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerDied","Data":"29f2e324112ee15af05481b82cd17d1f42324587a004b8d2f8b5d497d525b462"} Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.541327 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerDied","Data":"4b5da10bf892b403d3e86d53172428c255a8471861df08374e19429d62db2128"} Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.564617 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.5645936539999998 podStartE2EDuration="1.564593654s" podCreationTimestamp="2025-12-05 09:11:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:11:05.557312046 +0000 UTC m=+3330.045976668" watchObservedRunningTime="2025-12-05 09:11:05.564593654 +0000 UTC m=+3330.053258276" Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.566353 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" podUID="17ad89b3-21e5-4db2-8d47-29248aa9e054" Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.571047 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" podUID="17ad89b3-21e5-4db2-8d47-29248aa9e054" Dec 05 09:11:05 crc kubenswrapper[4876]: I1205 09:11:05.867870 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f" path="/var/lib/kubelet/pods/ffb0d6c6-bc3c-403d-ac74-bdc0b06c632f/volumes" Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.586518 4876 generic.go:334] "Generic (PLEG): container finished" podID="4601390b-7a61-460d-8f54-83886291c0c2" containerID="95c5aff4c7ebdbb3bfc46076190b6cbd4739b6d84e3f7a3dfb0022e996561207" exitCode=0 Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.586674 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerDied","Data":"95c5aff4c7ebdbb3bfc46076190b6cbd4739b6d84e3f7a3dfb0022e996561207"} Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.593818 4876 generic.go:334] "Generic (PLEG): container finished" podID="bf04a5a6-de08-4983-8986-c0c949090943" containerID="c63b6e6c01b11dcf2df1d4790bc24afdbae1f0cde40168fb4e95ee8fc277a5b0" exitCode=137 Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.597476 4876 generic.go:334] "Generic (PLEG): container finished" podID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerID="96c0a945ba122d80746127a1d30427b8322d9a1676dab585fcfd5627935ed4b3" exitCode=0 Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.597503 4876 generic.go:334] "Generic (PLEG): container finished" podID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerID="ca4ddcc6807b9eb6d5fa6e5b73997f5428e0f1b481da964ed040ee650c089911" exitCode=0 Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.597512 4876 generic.go:334] "Generic (PLEG): container finished" podID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerID="9e3064522f5bf0b86fc352814475b6b5bd89ce900d8725567559ba6497ea074f" exitCode=0 Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.597554 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerDied","Data":"96c0a945ba122d80746127a1d30427b8322d9a1676dab585fcfd5627935ed4b3"} Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.597592 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerDied","Data":"ca4ddcc6807b9eb6d5fa6e5b73997f5428e0f1b481da964ed040ee650c089911"} Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.597602 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerDied","Data":"9e3064522f5bf0b86fc352814475b6b5bd89ce900d8725567559ba6497ea074f"} Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.721249 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.724184 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf04a5a6-de08-4983-8986-c0c949090943" podUID="17ad89b3-21e5-4db2-8d47-29248aa9e054" Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.815681 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-combined-ca-bundle\") pod \"bf04a5a6-de08-4983-8986-c0c949090943\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.815814 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk98c\" (UniqueName: \"kubernetes.io/projected/bf04a5a6-de08-4983-8986-c0c949090943-kube-api-access-fk98c\") pod \"bf04a5a6-de08-4983-8986-c0c949090943\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.815937 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config-secret\") pod \"bf04a5a6-de08-4983-8986-c0c949090943\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " Dec 05 09:11:06 crc kubenswrapper[4876]: I1205 09:11:06.816535 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config\") pod \"bf04a5a6-de08-4983-8986-c0c949090943\" (UID: \"bf04a5a6-de08-4983-8986-c0c949090943\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:06.835519 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf04a5a6-de08-4983-8986-c0c949090943-kube-api-access-fk98c" (OuterVolumeSpecName: "kube-api-access-fk98c") pod "bf04a5a6-de08-4983-8986-c0c949090943" (UID: "bf04a5a6-de08-4983-8986-c0c949090943"). InnerVolumeSpecName "kube-api-access-fk98c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:06.850044 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bf04a5a6-de08-4983-8986-c0c949090943" (UID: "bf04a5a6-de08-4983-8986-c0c949090943"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:06.858136 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf04a5a6-de08-4983-8986-c0c949090943" (UID: "bf04a5a6-de08-4983-8986-c0c949090943"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:06.901040 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bf04a5a6-de08-4983-8986-c0c949090943" (UID: "bf04a5a6-de08-4983-8986-c0c949090943"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:06.919132 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:06.919169 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk98c\" (UniqueName: \"kubernetes.io/projected/bf04a5a6-de08-4983-8986-c0c949090943-kube-api-access-fk98c\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:06.919188 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:06.919199 4876 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bf04a5a6-de08-4983-8986-c0c949090943-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.009596 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.009987 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4sj77"] Dec 05 09:11:07 crc kubenswrapper[4876]: E1205 09:11:07.010383 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="init-config-reloader" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.010400 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="init-config-reloader" Dec 05 09:11:07 crc kubenswrapper[4876]: E1205 09:11:07.010439 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="config-reloader" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.010448 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="config-reloader" Dec 05 09:11:07 crc kubenswrapper[4876]: E1205 09:11:07.010461 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="thanos-sidecar" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.010469 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="thanos-sidecar" Dec 05 09:11:07 crc kubenswrapper[4876]: E1205 09:11:07.010488 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="prometheus" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.010496 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="prometheus" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.010741 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="config-reloader" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.010772 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="thanos-sidecar" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.010782 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" containerName="prometheus" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.012829 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.024251 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4sj77"] Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122085 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config\") pod \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122151 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config-out\") pod \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122209 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-tls-assets\") pod \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122235 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122264 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-web-config\") pod \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122317 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kql6m\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-kube-api-access-kql6m\") pod \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122396 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-prometheus-metric-storage-rulefiles-0\") pod \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122416 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-thanos-prometheus-http-client-file\") pod \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\" (UID: \"3cb6f8ab-8fcc-4518-88e7-4f479a828d00\") " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122685 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkl2p\" (UniqueName: \"kubernetes.io/projected/062e963d-bece-4a6c-aa68-90648656b0d0-kube-api-access-zkl2p\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122723 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/062e963d-bece-4a6c-aa68-90648656b0d0-utilities\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.122823 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/062e963d-bece-4a6c-aa68-90648656b0d0-catalog-content\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.123053 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "3cb6f8ab-8fcc-4518-88e7-4f479a828d00" (UID: "3cb6f8ab-8fcc-4518-88e7-4f479a828d00"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.125443 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config-out" (OuterVolumeSpecName: "config-out") pod "3cb6f8ab-8fcc-4518-88e7-4f479a828d00" (UID: "3cb6f8ab-8fcc-4518-88e7-4f479a828d00"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.126482 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "3cb6f8ab-8fcc-4518-88e7-4f479a828d00" (UID: "3cb6f8ab-8fcc-4518-88e7-4f479a828d00"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.126944 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "3cb6f8ab-8fcc-4518-88e7-4f479a828d00" (UID: "3cb6f8ab-8fcc-4518-88e7-4f479a828d00"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.127052 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-kube-api-access-kql6m" (OuterVolumeSpecName: "kube-api-access-kql6m") pod "3cb6f8ab-8fcc-4518-88e7-4f479a828d00" (UID: "3cb6f8ab-8fcc-4518-88e7-4f479a828d00"). InnerVolumeSpecName "kube-api-access-kql6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.129179 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config" (OuterVolumeSpecName: "config") pod "3cb6f8ab-8fcc-4518-88e7-4f479a828d00" (UID: "3cb6f8ab-8fcc-4518-88e7-4f479a828d00"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.132867 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "3cb6f8ab-8fcc-4518-88e7-4f479a828d00" (UID: "3cb6f8ab-8fcc-4518-88e7-4f479a828d00"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.150004 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-web-config" (OuterVolumeSpecName: "web-config") pod "3cb6f8ab-8fcc-4518-88e7-4f479a828d00" (UID: "3cb6f8ab-8fcc-4518-88e7-4f479a828d00"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224327 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkl2p\" (UniqueName: \"kubernetes.io/projected/062e963d-bece-4a6c-aa68-90648656b0d0-kube-api-access-zkl2p\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224379 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/062e963d-bece-4a6c-aa68-90648656b0d0-utilities\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224446 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/062e963d-bece-4a6c-aa68-90648656b0d0-catalog-content\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224554 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kql6m\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-kube-api-access-kql6m\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224567 4876 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224579 4876 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224590 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224599 4876 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-config-out\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224608 4876 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224658 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.224668 4876 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3cb6f8ab-8fcc-4518-88e7-4f479a828d00-web-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.225051 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/062e963d-bece-4a6c-aa68-90648656b0d0-utilities\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.225208 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/062e963d-bece-4a6c-aa68-90648656b0d0-catalog-content\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.242334 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkl2p\" (UniqueName: \"kubernetes.io/projected/062e963d-bece-4a6c-aa68-90648656b0d0-kube-api-access-zkl2p\") pod \"community-operators-4sj77\" (UID: \"062e963d-bece-4a6c-aa68-90648656b0d0\") " pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.244962 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.327067 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.346515 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.609696 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3cb6f8ab-8fcc-4518-88e7-4f479a828d00","Type":"ContainerDied","Data":"68b13db6025efff889a843e12fbda3344f880ccd77527af31cae918d9a4fa9b2"} Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.609802 4876 scope.go:117] "RemoveContainer" containerID="96c0a945ba122d80746127a1d30427b8322d9a1676dab585fcfd5627935ed4b3" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.609726 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.612287 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.630651 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf04a5a6-de08-4983-8986-c0c949090943" podUID="17ad89b3-21e5-4db2-8d47-29248aa9e054" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.679260 4876 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bf04a5a6-de08-4983-8986-c0c949090943" podUID="17ad89b3-21e5-4db2-8d47-29248aa9e054" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.685365 4876 scope.go:117] "RemoveContainer" containerID="ca4ddcc6807b9eb6d5fa6e5b73997f5428e0f1b481da964ed040ee650c089911" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.709831 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.724270 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.729729 4876 scope.go:117] "RemoveContainer" containerID="9e3064522f5bf0b86fc352814475b6b5bd89ce900d8725567559ba6497ea074f" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.735426 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.755734 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.755871 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.759107 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.759395 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.759883 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.759976 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.760484 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.769265 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-zx6qt" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.774679 4876 scope.go:117] "RemoveContainer" containerID="08f3d78e7146f8ec22955491551da094db8aee633e3449dcd3a46682ebbc1ebc" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.781430 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.810802 4876 scope.go:117] "RemoveContainer" containerID="c63b6e6c01b11dcf2df1d4790bc24afdbae1f0cde40168fb4e95ee8fc277a5b0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837113 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837155 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837207 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837232 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837264 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/516d1a02-6b2f-49af-ad71-75d10034e91e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837353 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-config\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837382 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv447\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-kube-api-access-gv447\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837405 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/516d1a02-6b2f-49af-ad71-75d10034e91e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837446 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837495 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.837548 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.873570 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb6f8ab-8fcc-4518-88e7-4f479a828d00" path="/var/lib/kubelet/pods/3cb6f8ab-8fcc-4518-88e7-4f479a828d00/volumes" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.875034 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf04a5a6-de08-4983-8986-c0c949090943" path="/var/lib/kubelet/pods/bf04a5a6-de08-4983-8986-c0c949090943/volumes" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.939079 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.939122 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.939339 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.939371 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/516d1a02-6b2f-49af-ad71-75d10034e91e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.939534 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-config\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.939657 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv447\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-kube-api-access-gv447\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.940142 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/516d1a02-6b2f-49af-ad71-75d10034e91e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.940770 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/516d1a02-6b2f-49af-ad71-75d10034e91e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.940849 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.941395 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.941482 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.941532 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.941587 4876 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.946202 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.947728 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.948886 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-config\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.951527 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.965830 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.968606 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.969347 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv447\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-kube-api-access-gv447\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.970491 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:07 crc kubenswrapper[4876]: I1205 09:11:07.974041 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/516d1a02-6b2f-49af-ad71-75d10034e91e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.006718 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"prometheus-metric-storage-0\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.088773 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.373253 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4sj77"] Dec 05 09:11:08 crc kubenswrapper[4876]: W1205 09:11:08.382484 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod062e963d_bece_4a6c_aa68_90648656b0d0.slice/crio-ae4def8e8556de52b435d72dcdbb77461f456b5ffd9f1f51dff43ded6b9acbf4 WatchSource:0}: Error finding container ae4def8e8556de52b435d72dcdbb77461f456b5ffd9f1f51dff43ded6b9acbf4: Status 404 returned error can't find the container with id ae4def8e8556de52b435d72dcdbb77461f456b5ffd9f1f51dff43ded6b9acbf4 Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.604399 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.637063 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4sj77" event={"ID":"062e963d-bece-4a6c-aa68-90648656b0d0","Type":"ContainerStarted","Data":"339cb004faf9cd589c1d832c01a1eab25272aa5e035dec40ef102a8a0e87a1f3"} Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.637252 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4sj77" event={"ID":"062e963d-bece-4a6c-aa68-90648656b0d0","Type":"ContainerStarted","Data":"ae4def8e8556de52b435d72dcdbb77461f456b5ffd9f1f51dff43ded6b9acbf4"} Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.657930 4876 generic.go:334] "Generic (PLEG): container finished" podID="4601390b-7a61-460d-8f54-83886291c0c2" containerID="4aafdc439d48b94b197c3f1b87d0c22892917fbd0a9fb609fc0d84730b8c99db" exitCode=0 Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.657971 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerDied","Data":"4aafdc439d48b94b197c3f1b87d0c22892917fbd0a9fb609fc0d84730b8c99db"} Dec 05 09:11:08 crc kubenswrapper[4876]: I1205 09:11:08.987889 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.179548 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-internal-tls-certs\") pod \"4601390b-7a61-460d-8f54-83886291c0c2\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.179626 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-public-tls-certs\") pod \"4601390b-7a61-460d-8f54-83886291c0c2\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.179839 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5mkc\" (UniqueName: \"kubernetes.io/projected/4601390b-7a61-460d-8f54-83886291c0c2-kube-api-access-q5mkc\") pod \"4601390b-7a61-460d-8f54-83886291c0c2\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.179996 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-config-data\") pod \"4601390b-7a61-460d-8f54-83886291c0c2\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.180106 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-scripts\") pod \"4601390b-7a61-460d-8f54-83886291c0c2\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.180371 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-combined-ca-bundle\") pod \"4601390b-7a61-460d-8f54-83886291c0c2\" (UID: \"4601390b-7a61-460d-8f54-83886291c0c2\") " Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.192184 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-scripts" (OuterVolumeSpecName: "scripts") pod "4601390b-7a61-460d-8f54-83886291c0c2" (UID: "4601390b-7a61-460d-8f54-83886291c0c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.192187 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4601390b-7a61-460d-8f54-83886291c0c2-kube-api-access-q5mkc" (OuterVolumeSpecName: "kube-api-access-q5mkc") pod "4601390b-7a61-460d-8f54-83886291c0c2" (UID: "4601390b-7a61-460d-8f54-83886291c0c2"). InnerVolumeSpecName "kube-api-access-q5mkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.284418 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4601390b-7a61-460d-8f54-83886291c0c2" (UID: "4601390b-7a61-460d-8f54-83886291c0c2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.287114 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.287250 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.287329 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5mkc\" (UniqueName: \"kubernetes.io/projected/4601390b-7a61-460d-8f54-83886291c0c2-kube-api-access-q5mkc\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.295468 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4601390b-7a61-460d-8f54-83886291c0c2" (UID: "4601390b-7a61-460d-8f54-83886291c0c2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.324367 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4601390b-7a61-460d-8f54-83886291c0c2" (UID: "4601390b-7a61-460d-8f54-83886291c0c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.373865 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-config-data" (OuterVolumeSpecName: "config-data") pod "4601390b-7a61-460d-8f54-83886291c0c2" (UID: "4601390b-7a61-460d-8f54-83886291c0c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.389259 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.389299 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.389312 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4601390b-7a61-460d-8f54-83886291c0c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.668851 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4601390b-7a61-460d-8f54-83886291c0c2","Type":"ContainerDied","Data":"5fb15807ccc9d4467fcee9f436056fb4c4246d9ea870deabd1543613009c861c"} Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.668889 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.668927 4876 scope.go:117] "RemoveContainer" containerID="95c5aff4c7ebdbb3bfc46076190b6cbd4739b6d84e3f7a3dfb0022e996561207" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.670746 4876 generic.go:334] "Generic (PLEG): container finished" podID="062e963d-bece-4a6c-aa68-90648656b0d0" containerID="339cb004faf9cd589c1d832c01a1eab25272aa5e035dec40ef102a8a0e87a1f3" exitCode=0 Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.670771 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4sj77" event={"ID":"062e963d-bece-4a6c-aa68-90648656b0d0","Type":"ContainerDied","Data":"339cb004faf9cd589c1d832c01a1eab25272aa5e035dec40ef102a8a0e87a1f3"} Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.672475 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerStarted","Data":"e230358130f28b56e727f86c839e9a7ec0d9cf0342dc650c09a865cadd798dc7"} Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.702999 4876 scope.go:117] "RemoveContainer" containerID="4aafdc439d48b94b197c3f1b87d0c22892917fbd0a9fb609fc0d84730b8c99db" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.727928 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.739545 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.745877 4876 scope.go:117] "RemoveContainer" containerID="29f2e324112ee15af05481b82cd17d1f42324587a004b8d2f8b5d497d525b462" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.757980 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 05 09:11:09 crc kubenswrapper[4876]: E1205 09:11:09.758445 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-evaluator" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.758465 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-evaluator" Dec 05 09:11:09 crc kubenswrapper[4876]: E1205 09:11:09.758482 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-api" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.758488 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-api" Dec 05 09:11:09 crc kubenswrapper[4876]: E1205 09:11:09.758499 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-listener" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.758505 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-listener" Dec 05 09:11:09 crc kubenswrapper[4876]: E1205 09:11:09.758527 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-notifier" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.758533 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-notifier" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.758727 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-listener" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.758740 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-api" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.758749 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-notifier" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.758763 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4601390b-7a61-460d-8f54-83886291c0c2" containerName="aodh-evaluator" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.760578 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.766494 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.766541 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.766712 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.766887 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tqf85" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.767053 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.771301 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.774414 4876 scope.go:117] "RemoveContainer" containerID="4b5da10bf892b403d3e86d53172428c255a8471861df08374e19429d62db2128" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.795539 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.795613 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-config-data\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.795732 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-scripts\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.795776 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4rw5\" (UniqueName: \"kubernetes.io/projected/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-kube-api-access-d4rw5\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.795814 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-internal-tls-certs\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.795841 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-public-tls-certs\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.835527 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4601390b-7a61-460d-8f54-83886291c0c2" path="/var/lib/kubelet/pods/4601390b-7a61-460d-8f54-83886291c0c2/volumes" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.897326 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-scripts\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.897426 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4rw5\" (UniqueName: \"kubernetes.io/projected/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-kube-api-access-d4rw5\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.897490 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-internal-tls-certs\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.897565 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-public-tls-certs\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.897642 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.897681 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-config-data\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.902789 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-internal-tls-certs\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.903874 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.904178 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-public-tls-certs\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.907436 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-scripts\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.913579 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-config-data\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:09 crc kubenswrapper[4876]: I1205 09:11:09.921173 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4rw5\" (UniqueName: \"kubernetes.io/projected/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-kube-api-access-d4rw5\") pod \"aodh-0\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " pod="openstack/aodh-0" Dec 05 09:11:10 crc kubenswrapper[4876]: I1205 09:11:10.081860 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:11:10 crc kubenswrapper[4876]: I1205 09:11:10.577915 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:11:10 crc kubenswrapper[4876]: W1205 09:11:10.584892 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd32d3b35_7ba5_40c9_a3b5_64edc93079c5.slice/crio-3169e3e7d16837ea5a3083f8d806891073fd17ac7feab59e1e549ab276ab6618 WatchSource:0}: Error finding container 3169e3e7d16837ea5a3083f8d806891073fd17ac7feab59e1e549ab276ab6618: Status 404 returned error can't find the container with id 3169e3e7d16837ea5a3083f8d806891073fd17ac7feab59e1e549ab276ab6618 Dec 05 09:11:10 crc kubenswrapper[4876]: I1205 09:11:10.689805 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerStarted","Data":"3169e3e7d16837ea5a3083f8d806891073fd17ac7feab59e1e549ab276ab6618"} Dec 05 09:11:11 crc kubenswrapper[4876]: I1205 09:11:11.704657 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerStarted","Data":"4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8"} Dec 05 09:11:12 crc kubenswrapper[4876]: I1205 09:11:12.720773 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerStarted","Data":"d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf"} Dec 05 09:11:12 crc kubenswrapper[4876]: I1205 09:11:12.723402 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerStarted","Data":"ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5"} Dec 05 09:11:15 crc kubenswrapper[4876]: I1205 09:11:15.760429 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4sj77" event={"ID":"062e963d-bece-4a6c-aa68-90648656b0d0","Type":"ContainerStarted","Data":"e1c5519e820c9837d5692da35a726a935235d1bb5d0eda4f40401c524441f94f"} Dec 05 09:11:15 crc kubenswrapper[4876]: I1205 09:11:15.765411 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerStarted","Data":"8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6"} Dec 05 09:11:16 crc kubenswrapper[4876]: I1205 09:11:16.801124 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerStarted","Data":"85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c"} Dec 05 09:11:16 crc kubenswrapper[4876]: I1205 09:11:16.805379 4876 generic.go:334] "Generic (PLEG): container finished" podID="062e963d-bece-4a6c-aa68-90648656b0d0" containerID="e1c5519e820c9837d5692da35a726a935235d1bb5d0eda4f40401c524441f94f" exitCode=0 Dec 05 09:11:16 crc kubenswrapper[4876]: I1205 09:11:16.806081 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4sj77" event={"ID":"062e963d-bece-4a6c-aa68-90648656b0d0","Type":"ContainerDied","Data":"e1c5519e820c9837d5692da35a726a935235d1bb5d0eda4f40401c524441f94f"} Dec 05 09:11:16 crc kubenswrapper[4876]: I1205 09:11:16.830383 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.9984534969999999 podStartE2EDuration="7.830360408s" podCreationTimestamp="2025-12-05 09:11:09 +0000 UTC" firstStartedPulling="2025-12-05 09:11:10.587744983 +0000 UTC m=+3335.076409595" lastFinishedPulling="2025-12-05 09:11:16.419651884 +0000 UTC m=+3340.908316506" observedRunningTime="2025-12-05 09:11:16.821346363 +0000 UTC m=+3341.310010995" watchObservedRunningTime="2025-12-05 09:11:16.830360408 +0000 UTC m=+3341.319025030" Dec 05 09:11:17 crc kubenswrapper[4876]: I1205 09:11:17.820404 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4sj77" event={"ID":"062e963d-bece-4a6c-aa68-90648656b0d0","Type":"ContainerStarted","Data":"ff8fc229b2a2fe98698b6689590d7ba699e1bcea38700eecc63989e425f55b05"} Dec 05 09:11:17 crc kubenswrapper[4876]: I1205 09:11:17.841893 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4sj77" podStartSLOduration=4.249797083 podStartE2EDuration="11.841869442s" podCreationTimestamp="2025-12-05 09:11:06 +0000 UTC" firstStartedPulling="2025-12-05 09:11:09.672297385 +0000 UTC m=+3334.160962007" lastFinishedPulling="2025-12-05 09:11:17.264369744 +0000 UTC m=+3341.753034366" observedRunningTime="2025-12-05 09:11:17.838213033 +0000 UTC m=+3342.326877655" watchObservedRunningTime="2025-12-05 09:11:17.841869442 +0000 UTC m=+3342.330534074" Dec 05 09:11:19 crc kubenswrapper[4876]: I1205 09:11:19.843401 4876 generic.go:334] "Generic (PLEG): container finished" podID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerID="ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5" exitCode=0 Dec 05 09:11:19 crc kubenswrapper[4876]: I1205 09:11:19.843530 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerDied","Data":"ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5"} Dec 05 09:11:21 crc kubenswrapper[4876]: I1205 09:11:21.867457 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerStarted","Data":"20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d"} Dec 05 09:11:24 crc kubenswrapper[4876]: I1205 09:11:24.902380 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerStarted","Data":"f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e"} Dec 05 09:11:26 crc kubenswrapper[4876]: I1205 09:11:26.928550 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerStarted","Data":"7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d"} Dec 05 09:11:26 crc kubenswrapper[4876]: I1205 09:11:26.954494 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=19.954473424 podStartE2EDuration="19.954473424s" podCreationTimestamp="2025-12-05 09:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:11:26.952286895 +0000 UTC m=+3351.440951527" watchObservedRunningTime="2025-12-05 09:11:26.954473424 +0000 UTC m=+3351.443138046" Dec 05 09:11:27 crc kubenswrapper[4876]: I1205 09:11:27.347956 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:27 crc kubenswrapper[4876]: I1205 09:11:27.348153 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:27 crc kubenswrapper[4876]: I1205 09:11:27.421668 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.003052 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4sj77" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.065608 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4sj77"] Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.090475 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.104116 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rlwz2"] Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.104376 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rlwz2" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerName="registry-server" containerID="cri-o://3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb" gracePeriod=2 Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.589010 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlwz2" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.768541 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-utilities\") pod \"c58822ac-9c07-4118-8743-8ace4b299aa1\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.768786 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sqs5\" (UniqueName: \"kubernetes.io/projected/c58822ac-9c07-4118-8743-8ace4b299aa1-kube-api-access-6sqs5\") pod \"c58822ac-9c07-4118-8743-8ace4b299aa1\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.768940 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-catalog-content\") pod \"c58822ac-9c07-4118-8743-8ace4b299aa1\" (UID: \"c58822ac-9c07-4118-8743-8ace4b299aa1\") " Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.770555 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-utilities" (OuterVolumeSpecName: "utilities") pod "c58822ac-9c07-4118-8743-8ace4b299aa1" (UID: "c58822ac-9c07-4118-8743-8ace4b299aa1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.779622 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58822ac-9c07-4118-8743-8ace4b299aa1-kube-api-access-6sqs5" (OuterVolumeSpecName: "kube-api-access-6sqs5") pod "c58822ac-9c07-4118-8743-8ace4b299aa1" (UID: "c58822ac-9c07-4118-8743-8ace4b299aa1"). InnerVolumeSpecName "kube-api-access-6sqs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.844802 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c58822ac-9c07-4118-8743-8ace4b299aa1" (UID: "c58822ac-9c07-4118-8743-8ace4b299aa1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.871649 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sqs5\" (UniqueName: \"kubernetes.io/projected/c58822ac-9c07-4118-8743-8ace4b299aa1-kube-api-access-6sqs5\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.871690 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.871705 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c58822ac-9c07-4118-8743-8ace4b299aa1-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.965787 4876 generic.go:334] "Generic (PLEG): container finished" podID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerID="3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb" exitCode=0 Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.966931 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlwz2" Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.977273 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlwz2" event={"ID":"c58822ac-9c07-4118-8743-8ace4b299aa1","Type":"ContainerDied","Data":"3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb"} Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.977347 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlwz2" event={"ID":"c58822ac-9c07-4118-8743-8ace4b299aa1","Type":"ContainerDied","Data":"96b09de799f5d821590d5b4b2213bf717b3f9ec027af6c3f7b26d0d95e4926d9"} Dec 05 09:11:28 crc kubenswrapper[4876]: I1205 09:11:28.977372 4876 scope.go:117] "RemoveContainer" containerID="3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.024911 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rlwz2"] Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.036490 4876 scope.go:117] "RemoveContainer" containerID="6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.044238 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rlwz2"] Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.067038 4876 scope.go:117] "RemoveContainer" containerID="0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.122344 4876 scope.go:117] "RemoveContainer" containerID="3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb" Dec 05 09:11:29 crc kubenswrapper[4876]: E1205 09:11:29.125697 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb\": container with ID starting with 3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb not found: ID does not exist" containerID="3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.125747 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb"} err="failed to get container status \"3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb\": rpc error: code = NotFound desc = could not find container \"3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb\": container with ID starting with 3a691e5a1218950a5c286268bd2139f119bc1cdba0712814db2580c582e1eddb not found: ID does not exist" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.125779 4876 scope.go:117] "RemoveContainer" containerID="6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4" Dec 05 09:11:29 crc kubenswrapper[4876]: E1205 09:11:29.126180 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4\": container with ID starting with 6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4 not found: ID does not exist" containerID="6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.126209 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4"} err="failed to get container status \"6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4\": rpc error: code = NotFound desc = could not find container \"6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4\": container with ID starting with 6b45f50d80ec78617c5bbb8510ba60b56f0e90134f285c9f665a1067b6d164d4 not found: ID does not exist" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.126227 4876 scope.go:117] "RemoveContainer" containerID="0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c" Dec 05 09:11:29 crc kubenswrapper[4876]: E1205 09:11:29.126521 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c\": container with ID starting with 0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c not found: ID does not exist" containerID="0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.126566 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c"} err="failed to get container status \"0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c\": rpc error: code = NotFound desc = could not find container \"0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c\": container with ID starting with 0f179e5bf57f770df9059c22ae70fdada9b47265630db3428166ca115d24c43c not found: ID does not exist" Dec 05 09:11:29 crc kubenswrapper[4876]: I1205 09:11:29.837435 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" path="/var/lib/kubelet/pods/c58822ac-9c07-4118-8743-8ace4b299aa1/volumes" Dec 05 09:11:38 crc kubenswrapper[4876]: I1205 09:11:38.090680 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:38 crc kubenswrapper[4876]: I1205 09:11:38.096568 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:38 crc kubenswrapper[4876]: I1205 09:11:38.213537 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:11:38 crc kubenswrapper[4876]: I1205 09:11:38.213597 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:11:39 crc kubenswrapper[4876]: I1205 09:11:39.077981 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.285673 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-94nw9"] Dec 05 09:11:41 crc kubenswrapper[4876]: E1205 09:11:41.286322 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerName="registry-server" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.286335 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerName="registry-server" Dec 05 09:11:41 crc kubenswrapper[4876]: E1205 09:11:41.286360 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerName="extract-utilities" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.286367 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerName="extract-utilities" Dec 05 09:11:41 crc kubenswrapper[4876]: E1205 09:11:41.286391 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerName="extract-content" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.286398 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerName="extract-content" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.286581 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="c58822ac-9c07-4118-8743-8ace4b299aa1" containerName="registry-server" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.288119 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.300162 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-94nw9"] Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.451298 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-utilities\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.451733 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz4jr\" (UniqueName: \"kubernetes.io/projected/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-kube-api-access-tz4jr\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.451938 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-catalog-content\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.553347 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-utilities\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.553419 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz4jr\" (UniqueName: \"kubernetes.io/projected/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-kube-api-access-tz4jr\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.553532 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-catalog-content\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.553888 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-utilities\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.553966 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-catalog-content\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.580249 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz4jr\" (UniqueName: \"kubernetes.io/projected/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-kube-api-access-tz4jr\") pod \"redhat-marketplace-94nw9\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:41 crc kubenswrapper[4876]: I1205 09:11:41.649336 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:42 crc kubenswrapper[4876]: I1205 09:11:42.218726 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-94nw9"] Dec 05 09:11:43 crc kubenswrapper[4876]: I1205 09:11:43.110082 4876 generic.go:334] "Generic (PLEG): container finished" podID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerID="b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b" exitCode=0 Dec 05 09:11:43 crc kubenswrapper[4876]: I1205 09:11:43.110193 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94nw9" event={"ID":"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce","Type":"ContainerDied","Data":"b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b"} Dec 05 09:11:43 crc kubenswrapper[4876]: I1205 09:11:43.110602 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94nw9" event={"ID":"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce","Type":"ContainerStarted","Data":"ffbe93121fdeb42e4a9780f5d35e88aed3dcac5648bec8d05e8aa81beed8db30"} Dec 05 09:11:44 crc kubenswrapper[4876]: I1205 09:11:44.122308 4876 generic.go:334] "Generic (PLEG): container finished" podID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerID="c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6" exitCode=0 Dec 05 09:11:44 crc kubenswrapper[4876]: I1205 09:11:44.122411 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94nw9" event={"ID":"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce","Type":"ContainerDied","Data":"c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6"} Dec 05 09:11:45 crc kubenswrapper[4876]: I1205 09:11:45.135245 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94nw9" event={"ID":"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce","Type":"ContainerStarted","Data":"84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f"} Dec 05 09:11:45 crc kubenswrapper[4876]: I1205 09:11:45.153777 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-94nw9" podStartSLOduration=2.618547425 podStartE2EDuration="4.153763008s" podCreationTimestamp="2025-12-05 09:11:41 +0000 UTC" firstStartedPulling="2025-12-05 09:11:43.112985138 +0000 UTC m=+3367.601649760" lastFinishedPulling="2025-12-05 09:11:44.648200711 +0000 UTC m=+3369.136865343" observedRunningTime="2025-12-05 09:11:45.150248542 +0000 UTC m=+3369.638913164" watchObservedRunningTime="2025-12-05 09:11:45.153763008 +0000 UTC m=+3369.642427630" Dec 05 09:11:51 crc kubenswrapper[4876]: I1205 09:11:51.650271 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:51 crc kubenswrapper[4876]: I1205 09:11:51.650884 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:51 crc kubenswrapper[4876]: I1205 09:11:51.700120 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:52 crc kubenswrapper[4876]: I1205 09:11:52.260321 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:52 crc kubenswrapper[4876]: I1205 09:11:52.314330 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-94nw9"] Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.233605 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-94nw9" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerName="registry-server" containerID="cri-o://84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f" gracePeriod=2 Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.732434 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.867796 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-utilities\") pod \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.868243 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz4jr\" (UniqueName: \"kubernetes.io/projected/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-kube-api-access-tz4jr\") pod \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.868357 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-catalog-content\") pod \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\" (UID: \"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce\") " Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.868764 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-utilities" (OuterVolumeSpecName: "utilities") pod "a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" (UID: "a28ba9b7-f8bb-4ce7-8463-b5f2043970ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.869243 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.874213 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-kube-api-access-tz4jr" (OuterVolumeSpecName: "kube-api-access-tz4jr") pod "a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" (UID: "a28ba9b7-f8bb-4ce7-8463-b5f2043970ce"). InnerVolumeSpecName "kube-api-access-tz4jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.886946 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" (UID: "a28ba9b7-f8bb-4ce7-8463-b5f2043970ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.971068 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz4jr\" (UniqueName: \"kubernetes.io/projected/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-kube-api-access-tz4jr\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:54 crc kubenswrapper[4876]: I1205 09:11:54.971112 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.246106 4876 generic.go:334] "Generic (PLEG): container finished" podID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerID="84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f" exitCode=0 Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.246153 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94nw9" event={"ID":"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce","Type":"ContainerDied","Data":"84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f"} Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.246182 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94nw9" event={"ID":"a28ba9b7-f8bb-4ce7-8463-b5f2043970ce","Type":"ContainerDied","Data":"ffbe93121fdeb42e4a9780f5d35e88aed3dcac5648bec8d05e8aa81beed8db30"} Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.246199 4876 scope.go:117] "RemoveContainer" containerID="84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.247305 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94nw9" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.273663 4876 scope.go:117] "RemoveContainer" containerID="c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.294141 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-94nw9"] Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.305483 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-94nw9"] Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.308522 4876 scope.go:117] "RemoveContainer" containerID="b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.346646 4876 scope.go:117] "RemoveContainer" containerID="84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f" Dec 05 09:11:55 crc kubenswrapper[4876]: E1205 09:11:55.347246 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f\": container with ID starting with 84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f not found: ID does not exist" containerID="84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.347304 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f"} err="failed to get container status \"84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f\": rpc error: code = NotFound desc = could not find container \"84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f\": container with ID starting with 84ea6a36c02a922da1a3ed95b40de651d9d62f5fe098c8470cd4897bac6efa5f not found: ID does not exist" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.347329 4876 scope.go:117] "RemoveContainer" containerID="c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6" Dec 05 09:11:55 crc kubenswrapper[4876]: E1205 09:11:55.347689 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6\": container with ID starting with c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6 not found: ID does not exist" containerID="c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.347716 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6"} err="failed to get container status \"c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6\": rpc error: code = NotFound desc = could not find container \"c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6\": container with ID starting with c18d25ba8d85339398521e86c1a63c547b0214903c7ed6fdb91f01d5472fcdf6 not found: ID does not exist" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.347735 4876 scope.go:117] "RemoveContainer" containerID="b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b" Dec 05 09:11:55 crc kubenswrapper[4876]: E1205 09:11:55.348045 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b\": container with ID starting with b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b not found: ID does not exist" containerID="b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.348081 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b"} err="failed to get container status \"b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b\": rpc error: code = NotFound desc = could not find container \"b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b\": container with ID starting with b819cc7d8fc7fe911a31f42b4ab74f78410f084d96414da29b8f9737c2b6976b not found: ID does not exist" Dec 05 09:11:55 crc kubenswrapper[4876]: I1205 09:11:55.839715 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" path="/var/lib/kubelet/pods/a28ba9b7-f8bb-4ce7-8463-b5f2043970ce/volumes" Dec 05 09:12:08 crc kubenswrapper[4876]: I1205 09:12:08.213921 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:12:08 crc kubenswrapper[4876]: I1205 09:12:08.214448 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.214029 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.214591 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.214640 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.215446 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df4e6443301eadd3c41cab90e72814aa0d48c3c05d4aabdd9044ffdf495e776d"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.215535 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://df4e6443301eadd3c41cab90e72814aa0d48c3c05d4aabdd9044ffdf495e776d" gracePeriod=600 Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.704924 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="df4e6443301eadd3c41cab90e72814aa0d48c3c05d4aabdd9044ffdf495e776d" exitCode=0 Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.704934 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"df4e6443301eadd3c41cab90e72814aa0d48c3c05d4aabdd9044ffdf495e776d"} Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.705264 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101"} Dec 05 09:12:38 crc kubenswrapper[4876]: I1205 09:12:38.705286 4876 scope.go:117] "RemoveContainer" containerID="21a1f587eada6d07133da0c5bfcd7893d3eb0b50d14a95c3e62f11eceaf71a8e" Dec 05 09:13:06 crc kubenswrapper[4876]: I1205 09:13:06.570354 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/manager/0.log" Dec 05 09:13:08 crc kubenswrapper[4876]: I1205 09:13:08.738221 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:13:08 crc kubenswrapper[4876]: I1205 09:13:08.738864 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="prometheus" containerID="cri-o://20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d" gracePeriod=600 Dec 05 09:13:08 crc kubenswrapper[4876]: I1205 09:13:08.739028 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="thanos-sidecar" containerID="cri-o://7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d" gracePeriod=600 Dec 05 09:13:08 crc kubenswrapper[4876]: I1205 09:13:08.739086 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="config-reloader" containerID="cri-o://f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e" gracePeriod=600 Dec 05 09:13:08 crc kubenswrapper[4876]: I1205 09:13:08.998159 4876 generic.go:334] "Generic (PLEG): container finished" podID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerID="7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d" exitCode=0 Dec 05 09:13:08 crc kubenswrapper[4876]: I1205 09:13:08.998491 4876 generic.go:334] "Generic (PLEG): container finished" podID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerID="20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d" exitCode=0 Dec 05 09:13:08 crc kubenswrapper[4876]: I1205 09:13:08.998383 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerDied","Data":"7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d"} Dec 05 09:13:08 crc kubenswrapper[4876]: I1205 09:13:08.998539 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerDied","Data":"20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d"} Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.704058 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780123 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780187 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-tls-assets\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780278 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/516d1a02-6b2f-49af-ad71-75d10034e91e-config-out\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780389 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-config\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780480 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780517 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gv447\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-kube-api-access-gv447\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780582 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780619 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-thanos-prometheus-http-client-file\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780669 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-secret-combined-ca-bundle\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780736 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.780765 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/516d1a02-6b2f-49af-ad71-75d10034e91e-prometheus-metric-storage-rulefiles-0\") pod \"516d1a02-6b2f-49af-ad71-75d10034e91e\" (UID: \"516d1a02-6b2f-49af-ad71-75d10034e91e\") " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.782038 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/516d1a02-6b2f-49af-ad71-75d10034e91e-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.792631 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-kube-api-access-gv447" (OuterVolumeSpecName: "kube-api-access-gv447") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "kube-api-access-gv447". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.812117 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/516d1a02-6b2f-49af-ad71-75d10034e91e-config-out" (OuterVolumeSpecName: "config-out") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.812277 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-config" (OuterVolumeSpecName: "config") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.813638 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.814286 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.815086 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.818475 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.826071 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.830396 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885593 4876 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885665 4876 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885694 4876 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885714 4876 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885732 4876 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/516d1a02-6b2f-49af-ad71-75d10034e91e-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885749 4876 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885768 4876 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885782 4876 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/516d1a02-6b2f-49af-ad71-75d10034e91e-config-out\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885798 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.885813 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gv447\" (UniqueName: \"kubernetes.io/projected/516d1a02-6b2f-49af-ad71-75d10034e91e-kube-api-access-gv447\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.951841 4876 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 05 09:13:09 crc kubenswrapper[4876]: I1205 09:13:09.987511 4876 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.034604 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config" (OuterVolumeSpecName: "web-config") pod "516d1a02-6b2f-49af-ad71-75d10034e91e" (UID: "516d1a02-6b2f-49af-ad71-75d10034e91e"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.036022 4876 generic.go:334] "Generic (PLEG): container finished" podID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerID="f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e" exitCode=0 Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.036102 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.036123 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerDied","Data":"f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e"} Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.036161 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"516d1a02-6b2f-49af-ad71-75d10034e91e","Type":"ContainerDied","Data":"e230358130f28b56e727f86c839e9a7ec0d9cf0342dc650c09a865cadd798dc7"} Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.036183 4876 scope.go:117] "RemoveContainer" containerID="7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.094176 4876 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/516d1a02-6b2f-49af-ad71-75d10034e91e-web-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.130616 4876 scope.go:117] "RemoveContainer" containerID="f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.162634 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.203863 4876 scope.go:117] "RemoveContainer" containerID="20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.204338 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.228153 4876 scope.go:117] "RemoveContainer" containerID="ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.256261 4876 scope.go:117] "RemoveContainer" containerID="7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d" Dec 05 09:13:10 crc kubenswrapper[4876]: E1205 09:13:10.257257 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d\": container with ID starting with 7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d not found: ID does not exist" containerID="7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.257314 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d"} err="failed to get container status \"7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d\": rpc error: code = NotFound desc = could not find container \"7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d\": container with ID starting with 7fba4ce8c8ea904183cc332731e6660cc55bf286b7aaf8592c883a526b36083d not found: ID does not exist" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.257349 4876 scope.go:117] "RemoveContainer" containerID="f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e" Dec 05 09:13:10 crc kubenswrapper[4876]: E1205 09:13:10.258023 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e\": container with ID starting with f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e not found: ID does not exist" containerID="f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.258078 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e"} err="failed to get container status \"f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e\": rpc error: code = NotFound desc = could not find container \"f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e\": container with ID starting with f55219d278cd7ce33c8ff993970d5ad1775179abcfca0214a642c883dee72e7e not found: ID does not exist" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.258113 4876 scope.go:117] "RemoveContainer" containerID="20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d" Dec 05 09:13:10 crc kubenswrapper[4876]: E1205 09:13:10.258829 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d\": container with ID starting with 20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d not found: ID does not exist" containerID="20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.258859 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d"} err="failed to get container status \"20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d\": rpc error: code = NotFound desc = could not find container \"20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d\": container with ID starting with 20ecfa2a977a4c7385058a7d9e0ff1dfc200c1622b1e413224753b5148f93b3d not found: ID does not exist" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.258877 4876 scope.go:117] "RemoveContainer" containerID="ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5" Dec 05 09:13:10 crc kubenswrapper[4876]: E1205 09:13:10.259365 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5\": container with ID starting with ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5 not found: ID does not exist" containerID="ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5" Dec 05 09:13:10 crc kubenswrapper[4876]: I1205 09:13:10.259392 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5"} err="failed to get container status \"ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5\": rpc error: code = NotFound desc = could not find container \"ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5\": container with ID starting with ca534c6c382f448e0e2629667f114bcf04debb4a278552fa54722cbf8050ddb5 not found: ID does not exist" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.031123 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:13:11 crc kubenswrapper[4876]: E1205 09:13:11.031583 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="thanos-sidecar" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.031605 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="thanos-sidecar" Dec 05 09:13:11 crc kubenswrapper[4876]: E1205 09:13:11.031624 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="prometheus" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.031631 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="prometheus" Dec 05 09:13:11 crc kubenswrapper[4876]: E1205 09:13:11.031645 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="config-reloader" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.031653 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="config-reloader" Dec 05 09:13:11 crc kubenswrapper[4876]: E1205 09:13:11.031671 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="init-config-reloader" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.031679 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="init-config-reloader" Dec 05 09:13:11 crc kubenswrapper[4876]: E1205 09:13:11.031695 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerName="extract-content" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.031702 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerName="extract-content" Dec 05 09:13:11 crc kubenswrapper[4876]: E1205 09:13:11.031724 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerName="registry-server" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.031731 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerName="registry-server" Dec 05 09:13:11 crc kubenswrapper[4876]: E1205 09:13:11.031741 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerName="extract-utilities" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.031749 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerName="extract-utilities" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.032013 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="prometheus" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.032037 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="a28ba9b7-f8bb-4ce7-8463-b5f2043970ce" containerName="registry-server" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.032049 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="config-reloader" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.037636 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" containerName="thanos-sidecar" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.042048 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.053622 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.053641 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.053855 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-zx6qt" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.053870 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.054170 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.054215 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.059008 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.059613 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.124771 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.124822 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.124861 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.125006 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.125086 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.125117 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.125136 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.125183 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.125245 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.125293 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.125321 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vgmj\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-kube-api-access-7vgmj\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226687 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226726 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226779 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226804 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226830 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226850 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226874 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226933 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.226987 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.227010 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vgmj\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-kube-api-access-7vgmj\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.227069 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.227853 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.227895 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.232302 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.234606 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.235128 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.235633 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.238256 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.238814 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.239395 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.243600 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.248652 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vgmj\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-kube-api-access-7vgmj\") pod \"prometheus-metric-storage-0\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.378475 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.837868 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="516d1a02-6b2f-49af-ad71-75d10034e91e" path="/var/lib/kubelet/pods/516d1a02-6b2f-49af-ad71-75d10034e91e/volumes" Dec 05 09:13:11 crc kubenswrapper[4876]: I1205 09:13:11.890349 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:13:12 crc kubenswrapper[4876]: I1205 09:13:12.061731 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerStarted","Data":"8671f05af281d1cf856a3c2199e45b63f4fc58c664ccf94383cadafc107b9103"} Dec 05 09:13:16 crc kubenswrapper[4876]: I1205 09:13:16.100595 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerStarted","Data":"e47885b757d3c3feb9638240324750284df0892a79a766ce82314d4e3006e64f"} Dec 05 09:13:24 crc kubenswrapper[4876]: I1205 09:13:24.211149 4876 generic.go:334] "Generic (PLEG): container finished" podID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerID="e47885b757d3c3feb9638240324750284df0892a79a766ce82314d4e3006e64f" exitCode=0 Dec 05 09:13:24 crc kubenswrapper[4876]: I1205 09:13:24.211264 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerDied","Data":"e47885b757d3c3feb9638240324750284df0892a79a766ce82314d4e3006e64f"} Dec 05 09:13:25 crc kubenswrapper[4876]: I1205 09:13:25.222557 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerStarted","Data":"f58a755ba1847244b7581364425921ba8d96ca7d7c0f9462a85c9095722e1c4c"} Dec 05 09:13:28 crc kubenswrapper[4876]: I1205 09:13:28.251433 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerStarted","Data":"94e3776fa89c98001e5a7d594be553ea97a042842a4297dd95cf24b108c47014"} Dec 05 09:13:29 crc kubenswrapper[4876]: I1205 09:13:29.262296 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerStarted","Data":"80a5784958231c9f3c7f8c6050a217940df22e3b641051e0cb4076ccd2c5ca28"} Dec 05 09:13:29 crc kubenswrapper[4876]: I1205 09:13:29.299858 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.299837844 podStartE2EDuration="18.299837844s" podCreationTimestamp="2025-12-05 09:13:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:13:29.287483679 +0000 UTC m=+3473.776148321" watchObservedRunningTime="2025-12-05 09:13:29.299837844 +0000 UTC m=+3473.788502466" Dec 05 09:13:31 crc kubenswrapper[4876]: I1205 09:13:31.379594 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:41 crc kubenswrapper[4876]: I1205 09:13:41.378726 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:41 crc kubenswrapper[4876]: I1205 09:13:41.384689 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:42 crc kubenswrapper[4876]: I1205 09:13:42.382305 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 05 09:13:43 crc kubenswrapper[4876]: I1205 09:13:43.809682 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kclrd"] Dec 05 09:13:43 crc kubenswrapper[4876]: I1205 09:13:43.813426 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:43 crc kubenswrapper[4876]: I1205 09:13:43.823192 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kclrd"] Dec 05 09:13:43 crc kubenswrapper[4876]: I1205 09:13:43.901022 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-catalog-content\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:43 crc kubenswrapper[4876]: I1205 09:13:43.901385 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4z8w\" (UniqueName: \"kubernetes.io/projected/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-kube-api-access-s4z8w\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:43 crc kubenswrapper[4876]: I1205 09:13:43.901528 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-utilities\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:44 crc kubenswrapper[4876]: I1205 09:13:44.003561 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4z8w\" (UniqueName: \"kubernetes.io/projected/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-kube-api-access-s4z8w\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:44 crc kubenswrapper[4876]: I1205 09:13:44.003838 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-utilities\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:44 crc kubenswrapper[4876]: I1205 09:13:44.004088 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-catalog-content\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:44 crc kubenswrapper[4876]: I1205 09:13:44.004296 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-utilities\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:44 crc kubenswrapper[4876]: I1205 09:13:44.004465 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-catalog-content\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:44 crc kubenswrapper[4876]: I1205 09:13:44.023176 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4z8w\" (UniqueName: \"kubernetes.io/projected/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-kube-api-access-s4z8w\") pod \"certified-operators-kclrd\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:44 crc kubenswrapper[4876]: I1205 09:13:44.144553 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:44 crc kubenswrapper[4876]: I1205 09:13:44.651608 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kclrd"] Dec 05 09:13:45 crc kubenswrapper[4876]: I1205 09:13:45.405444 4876 generic.go:334] "Generic (PLEG): container finished" podID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerID="1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8" exitCode=0 Dec 05 09:13:45 crc kubenswrapper[4876]: I1205 09:13:45.405539 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kclrd" event={"ID":"86b8f1a3-1edd-478c-ab68-e5d4a14a183c","Type":"ContainerDied","Data":"1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8"} Dec 05 09:13:45 crc kubenswrapper[4876]: I1205 09:13:45.405771 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kclrd" event={"ID":"86b8f1a3-1edd-478c-ab68-e5d4a14a183c","Type":"ContainerStarted","Data":"aa6318aa345f3a62c5927ee196c6af5f6cf60ed998d4bdd3ba3b5a9dc010d3fe"} Dec 05 09:13:46 crc kubenswrapper[4876]: I1205 09:13:46.415125 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kclrd" event={"ID":"86b8f1a3-1edd-478c-ab68-e5d4a14a183c","Type":"ContainerStarted","Data":"ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018"} Dec 05 09:13:48 crc kubenswrapper[4876]: I1205 09:13:48.435997 4876 generic.go:334] "Generic (PLEG): container finished" podID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerID="ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018" exitCode=0 Dec 05 09:13:48 crc kubenswrapper[4876]: I1205 09:13:48.436055 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kclrd" event={"ID":"86b8f1a3-1edd-478c-ab68-e5d4a14a183c","Type":"ContainerDied","Data":"ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018"} Dec 05 09:13:49 crc kubenswrapper[4876]: I1205 09:13:49.457015 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kclrd" event={"ID":"86b8f1a3-1edd-478c-ab68-e5d4a14a183c","Type":"ContainerStarted","Data":"c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7"} Dec 05 09:13:49 crc kubenswrapper[4876]: I1205 09:13:49.484047 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kclrd" podStartSLOduration=3.026191118 podStartE2EDuration="6.484026465s" podCreationTimestamp="2025-12-05 09:13:43 +0000 UTC" firstStartedPulling="2025-12-05 09:13:45.407810206 +0000 UTC m=+3489.896474828" lastFinishedPulling="2025-12-05 09:13:48.865645543 +0000 UTC m=+3493.354310175" observedRunningTime="2025-12-05 09:13:49.473627732 +0000 UTC m=+3493.962292354" watchObservedRunningTime="2025-12-05 09:13:49.484026465 +0000 UTC m=+3493.972691087" Dec 05 09:13:54 crc kubenswrapper[4876]: I1205 09:13:54.144878 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:54 crc kubenswrapper[4876]: I1205 09:13:54.145375 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:54 crc kubenswrapper[4876]: I1205 09:13:54.187117 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:54 crc kubenswrapper[4876]: I1205 09:13:54.553437 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:56 crc kubenswrapper[4876]: I1205 09:13:56.800757 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kclrd"] Dec 05 09:13:56 crc kubenswrapper[4876]: I1205 09:13:56.801313 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kclrd" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerName="registry-server" containerID="cri-o://c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7" gracePeriod=2 Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.286535 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.474937 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-utilities\") pod \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.475005 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-catalog-content\") pod \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.475156 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4z8w\" (UniqueName: \"kubernetes.io/projected/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-kube-api-access-s4z8w\") pod \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\" (UID: \"86b8f1a3-1edd-478c-ab68-e5d4a14a183c\") " Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.476019 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-utilities" (OuterVolumeSpecName: "utilities") pod "86b8f1a3-1edd-478c-ab68-e5d4a14a183c" (UID: "86b8f1a3-1edd-478c-ab68-e5d4a14a183c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.486140 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-kube-api-access-s4z8w" (OuterVolumeSpecName: "kube-api-access-s4z8w") pod "86b8f1a3-1edd-478c-ab68-e5d4a14a183c" (UID: "86b8f1a3-1edd-478c-ab68-e5d4a14a183c"). InnerVolumeSpecName "kube-api-access-s4z8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.525563 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86b8f1a3-1edd-478c-ab68-e5d4a14a183c" (UID: "86b8f1a3-1edd-478c-ab68-e5d4a14a183c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.534575 4876 generic.go:334] "Generic (PLEG): container finished" podID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerID="c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7" exitCode=0 Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.534614 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kclrd" event={"ID":"86b8f1a3-1edd-478c-ab68-e5d4a14a183c","Type":"ContainerDied","Data":"c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7"} Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.534641 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kclrd" event={"ID":"86b8f1a3-1edd-478c-ab68-e5d4a14a183c","Type":"ContainerDied","Data":"aa6318aa345f3a62c5927ee196c6af5f6cf60ed998d4bdd3ba3b5a9dc010d3fe"} Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.534658 4876 scope.go:117] "RemoveContainer" containerID="c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.534674 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kclrd" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.559883 4876 scope.go:117] "RemoveContainer" containerID="ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.587320 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.587374 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.587482 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4z8w\" (UniqueName: \"kubernetes.io/projected/86b8f1a3-1edd-478c-ab68-e5d4a14a183c-kube-api-access-s4z8w\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.598402 4876 scope.go:117] "RemoveContainer" containerID="1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.598527 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kclrd"] Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.608223 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kclrd"] Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.634009 4876 scope.go:117] "RemoveContainer" containerID="c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7" Dec 05 09:13:57 crc kubenswrapper[4876]: E1205 09:13:57.634511 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7\": container with ID starting with c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7 not found: ID does not exist" containerID="c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.634542 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7"} err="failed to get container status \"c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7\": rpc error: code = NotFound desc = could not find container \"c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7\": container with ID starting with c8c34d7868fc4117eb9d1388efe511d6ddb37c8ea8a4027814a5d6fa7940cfa7 not found: ID does not exist" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.634565 4876 scope.go:117] "RemoveContainer" containerID="ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018" Dec 05 09:13:57 crc kubenswrapper[4876]: E1205 09:13:57.634970 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018\": container with ID starting with ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018 not found: ID does not exist" containerID="ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.635019 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018"} err="failed to get container status \"ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018\": rpc error: code = NotFound desc = could not find container \"ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018\": container with ID starting with ca8b0e701883a1173e2331d67b10bd916cabc1bb4ef079807a33b072ab1e7018 not found: ID does not exist" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.635045 4876 scope.go:117] "RemoveContainer" containerID="1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8" Dec 05 09:13:57 crc kubenswrapper[4876]: E1205 09:13:57.635372 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8\": container with ID starting with 1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8 not found: ID does not exist" containerID="1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.635392 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8"} err="failed to get container status \"1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8\": rpc error: code = NotFound desc = could not find container \"1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8\": container with ID starting with 1adec0b082bf682126c4c72953b715367f9c0e3f4ccc3f31d910fc2aca9c35b8 not found: ID does not exist" Dec 05 09:13:57 crc kubenswrapper[4876]: I1205 09:13:57.834595 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" path="/var/lib/kubelet/pods/86b8f1a3-1edd-478c-ab68-e5d4a14a183c/volumes" Dec 05 09:14:38 crc kubenswrapper[4876]: I1205 09:14:38.213512 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:14:38 crc kubenswrapper[4876]: I1205 09:14:38.214267 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.182202 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh"] Dec 05 09:15:00 crc kubenswrapper[4876]: E1205 09:15:00.183114 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerName="registry-server" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.183128 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerName="registry-server" Dec 05 09:15:00 crc kubenswrapper[4876]: E1205 09:15:00.183171 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerName="extract-utilities" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.183178 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerName="extract-utilities" Dec 05 09:15:00 crc kubenswrapper[4876]: E1205 09:15:00.183192 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerName="extract-content" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.183198 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerName="extract-content" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.183414 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b8f1a3-1edd-478c-ab68-e5d4a14a183c" containerName="registry-server" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.185109 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.190416 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.190669 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.193234 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh"] Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.333533 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8tm9\" (UniqueName: \"kubernetes.io/projected/40d1d387-c945-4c38-b8d1-0b6c40025d78-kube-api-access-x8tm9\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.333633 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40d1d387-c945-4c38-b8d1-0b6c40025d78-config-volume\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.333696 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40d1d387-c945-4c38-b8d1-0b6c40025d78-secret-volume\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.435605 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8tm9\" (UniqueName: \"kubernetes.io/projected/40d1d387-c945-4c38-b8d1-0b6c40025d78-kube-api-access-x8tm9\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.436217 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40d1d387-c945-4c38-b8d1-0b6c40025d78-config-volume\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.437334 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40d1d387-c945-4c38-b8d1-0b6c40025d78-config-volume\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.437559 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40d1d387-c945-4c38-b8d1-0b6c40025d78-secret-volume\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.445265 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40d1d387-c945-4c38-b8d1-0b6c40025d78-secret-volume\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.460746 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8tm9\" (UniqueName: \"kubernetes.io/projected/40d1d387-c945-4c38-b8d1-0b6c40025d78-kube-api-access-x8tm9\") pod \"collect-profiles-29415435-7jksh\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.511176 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:00 crc kubenswrapper[4876]: I1205 09:15:00.968114 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh"] Dec 05 09:15:01 crc kubenswrapper[4876]: I1205 09:15:01.172815 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" event={"ID":"40d1d387-c945-4c38-b8d1-0b6c40025d78","Type":"ContainerStarted","Data":"69cc5ea1686330c3d704eccb2694bc34613c14bc9b9261882b8dc299cf754cd8"} Dec 05 09:15:01 crc kubenswrapper[4876]: I1205 09:15:01.172881 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" event={"ID":"40d1d387-c945-4c38-b8d1-0b6c40025d78","Type":"ContainerStarted","Data":"a25f58a89b66dae9e418c190c432a90d9e24ecb72a3f66c598c5de77d65e24f0"} Dec 05 09:15:01 crc kubenswrapper[4876]: I1205 09:15:01.191545 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" podStartSLOduration=1.191528979 podStartE2EDuration="1.191528979s" podCreationTimestamp="2025-12-05 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:15:01.187152229 +0000 UTC m=+3565.675816851" watchObservedRunningTime="2025-12-05 09:15:01.191528979 +0000 UTC m=+3565.680193601" Dec 05 09:15:02 crc kubenswrapper[4876]: I1205 09:15:02.185202 4876 generic.go:334] "Generic (PLEG): container finished" podID="40d1d387-c945-4c38-b8d1-0b6c40025d78" containerID="69cc5ea1686330c3d704eccb2694bc34613c14bc9b9261882b8dc299cf754cd8" exitCode=0 Dec 05 09:15:02 crc kubenswrapper[4876]: I1205 09:15:02.185277 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" event={"ID":"40d1d387-c945-4c38-b8d1-0b6c40025d78","Type":"ContainerDied","Data":"69cc5ea1686330c3d704eccb2694bc34613c14bc9b9261882b8dc299cf754cd8"} Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.558304 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.697823 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8tm9\" (UniqueName: \"kubernetes.io/projected/40d1d387-c945-4c38-b8d1-0b6c40025d78-kube-api-access-x8tm9\") pod \"40d1d387-c945-4c38-b8d1-0b6c40025d78\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.697926 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40d1d387-c945-4c38-b8d1-0b6c40025d78-config-volume\") pod \"40d1d387-c945-4c38-b8d1-0b6c40025d78\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.698233 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40d1d387-c945-4c38-b8d1-0b6c40025d78-secret-volume\") pod \"40d1d387-c945-4c38-b8d1-0b6c40025d78\" (UID: \"40d1d387-c945-4c38-b8d1-0b6c40025d78\") " Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.699622 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40d1d387-c945-4c38-b8d1-0b6c40025d78-config-volume" (OuterVolumeSpecName: "config-volume") pod "40d1d387-c945-4c38-b8d1-0b6c40025d78" (UID: "40d1d387-c945-4c38-b8d1-0b6c40025d78"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.705718 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40d1d387-c945-4c38-b8d1-0b6c40025d78-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "40d1d387-c945-4c38-b8d1-0b6c40025d78" (UID: "40d1d387-c945-4c38-b8d1-0b6c40025d78"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.705891 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40d1d387-c945-4c38-b8d1-0b6c40025d78-kube-api-access-x8tm9" (OuterVolumeSpecName: "kube-api-access-x8tm9") pod "40d1d387-c945-4c38-b8d1-0b6c40025d78" (UID: "40d1d387-c945-4c38-b8d1-0b6c40025d78"). InnerVolumeSpecName "kube-api-access-x8tm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.800522 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40d1d387-c945-4c38-b8d1-0b6c40025d78-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.800824 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8tm9\" (UniqueName: \"kubernetes.io/projected/40d1d387-c945-4c38-b8d1-0b6c40025d78-kube-api-access-x8tm9\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:03 crc kubenswrapper[4876]: I1205 09:15:03.800834 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40d1d387-c945-4c38-b8d1-0b6c40025d78-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:04 crc kubenswrapper[4876]: I1205 09:15:04.209396 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" Dec 05 09:15:04 crc kubenswrapper[4876]: I1205 09:15:04.209481 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-7jksh" event={"ID":"40d1d387-c945-4c38-b8d1-0b6c40025d78","Type":"ContainerDied","Data":"a25f58a89b66dae9e418c190c432a90d9e24ecb72a3f66c598c5de77d65e24f0"} Dec 05 09:15:04 crc kubenswrapper[4876]: I1205 09:15:04.209526 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a25f58a89b66dae9e418c190c432a90d9e24ecb72a3f66c598c5de77d65e24f0" Dec 05 09:15:04 crc kubenswrapper[4876]: I1205 09:15:04.305642 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm"] Dec 05 09:15:04 crc kubenswrapper[4876]: I1205 09:15:04.316849 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415390-hm9qm"] Dec 05 09:15:05 crc kubenswrapper[4876]: I1205 09:15:05.835505 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9" path="/var/lib/kubelet/pods/4aec6a5d-c8c9-4046-8fe8-b61d6345f0f9/volumes" Dec 05 09:15:08 crc kubenswrapper[4876]: I1205 09:15:08.104610 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-jljwh"] Dec 05 09:15:08 crc kubenswrapper[4876]: I1205 09:15:08.141325 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-jljwh"] Dec 05 09:15:08 crc kubenswrapper[4876]: I1205 09:15:08.214262 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:15:08 crc kubenswrapper[4876]: I1205 09:15:08.214319 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:15:08 crc kubenswrapper[4876]: I1205 09:15:08.977672 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/manager/0.log" Dec 05 09:15:09 crc kubenswrapper[4876]: I1205 09:15:09.039331 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-6a58-account-create-update-txbfr"] Dec 05 09:15:09 crc kubenswrapper[4876]: I1205 09:15:09.056880 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-6a58-account-create-update-txbfr"] Dec 05 09:15:09 crc kubenswrapper[4876]: I1205 09:15:09.838766 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="316c29a1-4346-4be0-be25-aa29a1a9dd0d" path="/var/lib/kubelet/pods/316c29a1-4346-4be0-be25-aa29a1a9dd0d/volumes" Dec 05 09:15:09 crc kubenswrapper[4876]: I1205 09:15:09.839846 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f71887-ac32-4542-9a91-d1e350331dbc" path="/var/lib/kubelet/pods/e0f71887-ac32-4542-9a91-d1e350331dbc/volumes" Dec 05 09:15:10 crc kubenswrapper[4876]: I1205 09:15:10.441223 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 05 09:15:10 crc kubenswrapper[4876]: I1205 09:15:10.441568 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-api" containerID="cri-o://4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8" gracePeriod=30 Dec 05 09:15:10 crc kubenswrapper[4876]: I1205 09:15:10.441622 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-listener" containerID="cri-o://85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c" gracePeriod=30 Dec 05 09:15:10 crc kubenswrapper[4876]: I1205 09:15:10.441719 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-notifier" containerID="cri-o://8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6" gracePeriod=30 Dec 05 09:15:10 crc kubenswrapper[4876]: I1205 09:15:10.443119 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-evaluator" containerID="cri-o://d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf" gracePeriod=30 Dec 05 09:15:11 crc kubenswrapper[4876]: I1205 09:15:11.270360 4876 generic.go:334] "Generic (PLEG): container finished" podID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerID="d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf" exitCode=0 Dec 05 09:15:11 crc kubenswrapper[4876]: I1205 09:15:11.270650 4876 generic.go:334] "Generic (PLEG): container finished" podID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerID="4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8" exitCode=0 Dec 05 09:15:11 crc kubenswrapper[4876]: I1205 09:15:11.270671 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerDied","Data":"d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf"} Dec 05 09:15:11 crc kubenswrapper[4876]: I1205 09:15:11.270695 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerDied","Data":"4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8"} Dec 05 09:15:13 crc kubenswrapper[4876]: I1205 09:15:13.301410 4876 generic.go:334] "Generic (PLEG): container finished" podID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerID="8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6" exitCode=0 Dec 05 09:15:13 crc kubenswrapper[4876]: I1205 09:15:13.301518 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerDied","Data":"8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6"} Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.290353 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.314483 4876 generic.go:334] "Generic (PLEG): container finished" podID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerID="85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c" exitCode=0 Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.314523 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerDied","Data":"85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c"} Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.314549 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d32d3b35-7ba5-40c9-a3b5-64edc93079c5","Type":"ContainerDied","Data":"3169e3e7d16837ea5a3083f8d806891073fd17ac7feab59e1e549ab276ab6618"} Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.314565 4876 scope.go:117] "RemoveContainer" containerID="85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.314691 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.344714 4876 scope.go:117] "RemoveContainer" containerID="8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.378753 4876 scope.go:117] "RemoveContainer" containerID="d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.399248 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-scripts\") pod \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.399331 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4rw5\" (UniqueName: \"kubernetes.io/projected/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-kube-api-access-d4rw5\") pod \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.399368 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-config-data\") pod \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.399458 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-internal-tls-certs\") pod \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.399550 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-combined-ca-bundle\") pod \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.399621 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-public-tls-certs\") pod \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\" (UID: \"d32d3b35-7ba5-40c9-a3b5-64edc93079c5\") " Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.405665 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-kube-api-access-d4rw5" (OuterVolumeSpecName: "kube-api-access-d4rw5") pod "d32d3b35-7ba5-40c9-a3b5-64edc93079c5" (UID: "d32d3b35-7ba5-40c9-a3b5-64edc93079c5"). InnerVolumeSpecName "kube-api-access-d4rw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.411646 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-scripts" (OuterVolumeSpecName: "scripts") pod "d32d3b35-7ba5-40c9-a3b5-64edc93079c5" (UID: "d32d3b35-7ba5-40c9-a3b5-64edc93079c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.414158 4876 scope.go:117] "RemoveContainer" containerID="4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.476345 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d32d3b35-7ba5-40c9-a3b5-64edc93079c5" (UID: "d32d3b35-7ba5-40c9-a3b5-64edc93079c5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.486157 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d32d3b35-7ba5-40c9-a3b5-64edc93079c5" (UID: "d32d3b35-7ba5-40c9-a3b5-64edc93079c5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.502212 4876 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.502243 4876 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.502256 4876 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.502265 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4rw5\" (UniqueName: \"kubernetes.io/projected/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-kube-api-access-d4rw5\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.518743 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d32d3b35-7ba5-40c9-a3b5-64edc93079c5" (UID: "d32d3b35-7ba5-40c9-a3b5-64edc93079c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.528977 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-config-data" (OuterVolumeSpecName: "config-data") pod "d32d3b35-7ba5-40c9-a3b5-64edc93079c5" (UID: "d32d3b35-7ba5-40c9-a3b5-64edc93079c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.544508 4876 scope.go:117] "RemoveContainer" containerID="85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c" Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.547729 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c\": container with ID starting with 85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c not found: ID does not exist" containerID="85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.547773 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c"} err="failed to get container status \"85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c\": rpc error: code = NotFound desc = could not find container \"85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c\": container with ID starting with 85d3094c0922c993d92bcf5b1394634236e2bc8f83eb4911aaebb3975da4157c not found: ID does not exist" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.547802 4876 scope.go:117] "RemoveContainer" containerID="8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6" Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.548149 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6\": container with ID starting with 8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6 not found: ID does not exist" containerID="8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.548232 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6"} err="failed to get container status \"8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6\": rpc error: code = NotFound desc = could not find container \"8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6\": container with ID starting with 8eca6c82772099e6daaee177ba05fc8ec5daa619e0a779177a13ace4bc5f8cf6 not found: ID does not exist" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.548297 4876 scope.go:117] "RemoveContainer" containerID="d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf" Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.548809 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf\": container with ID starting with d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf not found: ID does not exist" containerID="d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.548849 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf"} err="failed to get container status \"d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf\": rpc error: code = NotFound desc = could not find container \"d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf\": container with ID starting with d20c259a07d566a1e0a26c9a66b4b45bef1a2f4f403cf5d6b8695600dbca0ccf not found: ID does not exist" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.548866 4876 scope.go:117] "RemoveContainer" containerID="4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8" Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.549124 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8\": container with ID starting with 4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8 not found: ID does not exist" containerID="4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.549150 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8"} err="failed to get container status \"4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8\": rpc error: code = NotFound desc = could not find container \"4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8\": container with ID starting with 4a662fc3bcd1980827d7c3d7fb654609ec36a983fac39eca43a53ea76c5761e8 not found: ID does not exist" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.604208 4876 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.604243 4876 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32d3b35-7ba5-40c9-a3b5-64edc93079c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.644963 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.654320 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.675367 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.675863 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-api" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.675886 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-api" Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.675941 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-listener" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.675948 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-listener" Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.675964 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d1d387-c945-4c38-b8d1-0b6c40025d78" containerName="collect-profiles" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.675972 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d1d387-c945-4c38-b8d1-0b6c40025d78" containerName="collect-profiles" Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.675986 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-notifier" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.675993 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-notifier" Dec 05 09:15:14 crc kubenswrapper[4876]: E1205 09:15:14.676002 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-evaluator" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.676009 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-evaluator" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.676204 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-evaluator" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.676218 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-notifier" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.676232 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-listener" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.676242 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" containerName="aodh-api" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.676260 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="40d1d387-c945-4c38-b8d1-0b6c40025d78" containerName="collect-profiles" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.678062 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.679771 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.681149 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.681383 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.682586 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tqf85" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.683065 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.707836 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.806881 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c88gt\" (UniqueName: \"kubernetes.io/projected/e960eaad-8be1-47d2-8e07-a17915f9cfea-kube-api-access-c88gt\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.806936 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.807060 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-scripts\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.807099 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-config-data\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.807116 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-internal-tls-certs\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.807140 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-public-tls-certs\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.909162 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c88gt\" (UniqueName: \"kubernetes.io/projected/e960eaad-8be1-47d2-8e07-a17915f9cfea-kube-api-access-c88gt\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.909213 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.909405 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-scripts\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.909485 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-config-data\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.909532 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-internal-tls-certs\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.909575 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-public-tls-certs\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.915522 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.915558 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-internal-tls-certs\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.915999 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-public-tls-certs\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.916518 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-config-data\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.916548 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e960eaad-8be1-47d2-8e07-a17915f9cfea-scripts\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.926495 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c88gt\" (UniqueName: \"kubernetes.io/projected/e960eaad-8be1-47d2-8e07-a17915f9cfea-kube-api-access-c88gt\") pod \"aodh-0\" (UID: \"e960eaad-8be1-47d2-8e07-a17915f9cfea\") " pod="openstack/aodh-0" Dec 05 09:15:14 crc kubenswrapper[4876]: I1205 09:15:14.996034 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 05 09:15:15 crc kubenswrapper[4876]: I1205 09:15:15.443551 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 05 09:15:15 crc kubenswrapper[4876]: I1205 09:15:15.452374 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:15:15 crc kubenswrapper[4876]: I1205 09:15:15.837826 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d32d3b35-7ba5-40c9-a3b5-64edc93079c5" path="/var/lib/kubelet/pods/d32d3b35-7ba5-40c9-a3b5-64edc93079c5/volumes" Dec 05 09:15:16 crc kubenswrapper[4876]: I1205 09:15:16.339398 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e960eaad-8be1-47d2-8e07-a17915f9cfea","Type":"ContainerStarted","Data":"1d162e559c768879634658f58fcc478bbd572a2bb4b496006b5843c6bc1f4fb9"} Dec 05 09:15:16 crc kubenswrapper[4876]: I1205 09:15:16.339981 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e960eaad-8be1-47d2-8e07-a17915f9cfea","Type":"ContainerStarted","Data":"2e3481b159e9fc432ee82979e03e0e8f0a80c46218e5c5c8d0c1a34603a7a51c"} Dec 05 09:15:17 crc kubenswrapper[4876]: I1205 09:15:17.352755 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e960eaad-8be1-47d2-8e07-a17915f9cfea","Type":"ContainerStarted","Data":"b30da3961680b37080aa5c38ab6a8a5fea7a07714baca56d444c1bd8b1750d3d"} Dec 05 09:15:18 crc kubenswrapper[4876]: I1205 09:15:18.364165 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e960eaad-8be1-47d2-8e07-a17915f9cfea","Type":"ContainerStarted","Data":"7587f3c014a86eec9033134211de69e924410bf5291e49c7037e42368677c70a"} Dec 05 09:15:19 crc kubenswrapper[4876]: I1205 09:15:19.375824 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e960eaad-8be1-47d2-8e07-a17915f9cfea","Type":"ContainerStarted","Data":"ce668cfc8d74195fbf508292b25838496174fcac54b701fa7d30296658562dd0"} Dec 05 09:15:19 crc kubenswrapper[4876]: I1205 09:15:19.407150 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.460133059 podStartE2EDuration="5.40711146s" podCreationTimestamp="2025-12-05 09:15:14 +0000 UTC" firstStartedPulling="2025-12-05 09:15:15.452034127 +0000 UTC m=+3579.940698749" lastFinishedPulling="2025-12-05 09:15:18.399012528 +0000 UTC m=+3582.887677150" observedRunningTime="2025-12-05 09:15:19.399865942 +0000 UTC m=+3583.888530584" watchObservedRunningTime="2025-12-05 09:15:19.40711146 +0000 UTC m=+3583.895776152" Dec 05 09:15:20 crc kubenswrapper[4876]: I1205 09:15:20.035592 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-6wsbn"] Dec 05 09:15:20 crc kubenswrapper[4876]: I1205 09:15:20.051693 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-6wsbn"] Dec 05 09:15:21 crc kubenswrapper[4876]: I1205 09:15:21.836113 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52564344-c699-4703-bc3b-5170d5e891c7" path="/var/lib/kubelet/pods/52564344-c699-4703-bc3b-5170d5e891c7/volumes" Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.213425 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.214815 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.214957 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.215802 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.215989 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" gracePeriod=600 Dec 05 09:15:38 crc kubenswrapper[4876]: E1205 09:15:38.336008 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.596012 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" exitCode=0 Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.596051 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101"} Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.596102 4876 scope.go:117] "RemoveContainer" containerID="df4e6443301eadd3c41cab90e72814aa0d48c3c05d4aabdd9044ffdf495e776d" Dec 05 09:15:38 crc kubenswrapper[4876]: I1205 09:15:38.596800 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:15:38 crc kubenswrapper[4876]: E1205 09:15:38.597180 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:15:42 crc kubenswrapper[4876]: I1205 09:15:42.687076 4876 scope.go:117] "RemoveContainer" containerID="9bdc5e8bd2d68fceed99329241e11014fff9acb9a49aee40f89ff67098c3866a" Dec 05 09:15:42 crc kubenswrapper[4876]: I1205 09:15:42.719704 4876 scope.go:117] "RemoveContainer" containerID="e1af3f3a242a1506c4cdd1a68145e367df462320c7f7f1f1673d0f595b09dfa6" Dec 05 09:15:42 crc kubenswrapper[4876]: I1205 09:15:42.790279 4876 scope.go:117] "RemoveContainer" containerID="79e81b17f0124fd628a2e83af4ad9805c82f27c4cc5e189df2218ebe5c677100" Dec 05 09:15:42 crc kubenswrapper[4876]: I1205 09:15:42.840808 4876 scope.go:117] "RemoveContainer" containerID="ac3466d10336d513d8332d7902558d8e71c5e3bdb7bc711014e1eab9c5d57166" Dec 05 09:15:49 crc kubenswrapper[4876]: I1205 09:15:49.824104 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:15:49 crc kubenswrapper[4876]: E1205 09:15:49.824991 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:16:04 crc kubenswrapper[4876]: I1205 09:16:04.824446 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:16:04 crc kubenswrapper[4876]: E1205 09:16:04.825243 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:16:15 crc kubenswrapper[4876]: I1205 09:16:15.833272 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:16:15 crc kubenswrapper[4876]: E1205 09:16:15.834046 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:16:28 crc kubenswrapper[4876]: I1205 09:16:28.824220 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:16:28 crc kubenswrapper[4876]: E1205 09:16:28.825252 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:16:42 crc kubenswrapper[4876]: I1205 09:16:42.824049 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:16:42 crc kubenswrapper[4876]: E1205 09:16:42.825018 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:16:54 crc kubenswrapper[4876]: I1205 09:16:54.824023 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:16:54 crc kubenswrapper[4876]: E1205 09:16:54.824821 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:17:08 crc kubenswrapper[4876]: I1205 09:17:08.824398 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:17:08 crc kubenswrapper[4876]: E1205 09:17:08.825435 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:17:11 crc kubenswrapper[4876]: I1205 09:17:11.140642 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/manager/0.log" Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.055108 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.056162 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="config-reloader" containerID="cri-o://94e3776fa89c98001e5a7d594be553ea97a042842a4297dd95cf24b108c47014" gracePeriod=600 Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.056232 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="thanos-sidecar" containerID="cri-o://80a5784958231c9f3c7f8c6050a217940df22e3b641051e0cb4076ccd2c5ca28" gracePeriod=600 Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.056161 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="prometheus" containerID="cri-o://f58a755ba1847244b7581364425921ba8d96ca7d7c0f9462a85c9095722e1c4c" gracePeriod=600 Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.602432 4876 generic.go:334] "Generic (PLEG): container finished" podID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerID="80a5784958231c9f3c7f8c6050a217940df22e3b641051e0cb4076ccd2c5ca28" exitCode=0 Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.602754 4876 generic.go:334] "Generic (PLEG): container finished" podID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerID="94e3776fa89c98001e5a7d594be553ea97a042842a4297dd95cf24b108c47014" exitCode=0 Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.602767 4876 generic.go:334] "Generic (PLEG): container finished" podID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerID="f58a755ba1847244b7581364425921ba8d96ca7d7c0f9462a85c9095722e1c4c" exitCode=0 Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.602506 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerDied","Data":"80a5784958231c9f3c7f8c6050a217940df22e3b641051e0cb4076ccd2c5ca28"} Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.602827 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerDied","Data":"94e3776fa89c98001e5a7d594be553ea97a042842a4297dd95cf24b108c47014"} Dec 05 09:17:15 crc kubenswrapper[4876]: I1205 09:17:15.602848 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerDied","Data":"f58a755ba1847244b7581364425921ba8d96ca7d7c0f9462a85c9095722e1c4c"} Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.031107 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.177982 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-thanos-prometheus-http-client-file\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178392 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178482 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config-out\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178520 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178566 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178603 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vgmj\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-kube-api-access-7vgmj\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178633 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178705 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-rulefiles-0\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178747 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-tls-assets\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178806 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-secret-combined-ca-bundle\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.178938 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-db\") pod \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\" (UID: \"189edc37-9fe6-4b23-8ee4-02b72a4bb17d\") " Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.182828 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-db" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "prometheus-metric-storage-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.184614 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-kube-api-access-7vgmj" (OuterVolumeSpecName: "kube-api-access-7vgmj") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "kube-api-access-7vgmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.184785 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.186828 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.187978 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.189508 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.190558 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config-out" (OuterVolumeSpecName: "config-out") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.190795 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.190822 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config" (OuterVolumeSpecName: "config") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.205492 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.270156 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config" (OuterVolumeSpecName: "web-config") pod "189edc37-9fe6-4b23-8ee4-02b72a4bb17d" (UID: "189edc37-9fe6-4b23-8ee4-02b72a4bb17d"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281760 4876 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281802 4876 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config-out\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281816 4876 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281828 4876 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281841 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vgmj\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-kube-api-access-7vgmj\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281850 4876 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-web-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281861 4876 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281871 4876 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281881 4876 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281891 4876 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-prometheus-metric-storage-db\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.281916 4876 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/189edc37-9fe6-4b23-8ee4-02b72a4bb17d-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.612646 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"189edc37-9fe6-4b23-8ee4-02b72a4bb17d","Type":"ContainerDied","Data":"8671f05af281d1cf856a3c2199e45b63f4fc58c664ccf94383cadafc107b9103"} Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.612699 4876 scope.go:117] "RemoveContainer" containerID="80a5784958231c9f3c7f8c6050a217940df22e3b641051e0cb4076ccd2c5ca28" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.612714 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.645196 4876 scope.go:117] "RemoveContainer" containerID="94e3776fa89c98001e5a7d594be553ea97a042842a4297dd95cf24b108c47014" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.647430 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.657051 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.672955 4876 scope.go:117] "RemoveContainer" containerID="f58a755ba1847244b7581364425921ba8d96ca7d7c0f9462a85c9095722e1c4c" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.681743 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:17:16 crc kubenswrapper[4876]: E1205 09:17:16.682154 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="prometheus" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.682171 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="prometheus" Dec 05 09:17:16 crc kubenswrapper[4876]: E1205 09:17:16.682192 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="thanos-sidecar" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.682199 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="thanos-sidecar" Dec 05 09:17:16 crc kubenswrapper[4876]: E1205 09:17:16.682217 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="config-reloader" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.682223 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="config-reloader" Dec 05 09:17:16 crc kubenswrapper[4876]: E1205 09:17:16.682239 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="init-config-reloader" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.682245 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="init-config-reloader" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.682455 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="prometheus" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.682488 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="thanos-sidecar" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.682501 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" containerName="config-reloader" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.684268 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.692566 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.692590 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.692725 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.692803 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.696804 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.696960 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-zx6qt" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.698020 4876 scope.go:117] "RemoveContainer" containerID="e47885b757d3c3feb9638240324750284df0892a79a766ce82314d4e3006e64f" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.701668 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.726570 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.794530 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-config\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.794623 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.794675 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ff40adb3-f509-46ed-96f3-c915e75906b0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.794716 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ff40adb3-f509-46ed-96f3-c915e75906b0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.794757 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.794826 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.795220 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ff40adb3-f509-46ed-96f3-c915e75906b0-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.795267 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ff40adb3-f509-46ed-96f3-c915e75906b0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.795355 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.795402 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.795441 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8xhm\" (UniqueName: \"kubernetes.io/projected/ff40adb3-f509-46ed-96f3-c915e75906b0-kube-api-access-n8xhm\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.897597 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ff40adb3-f509-46ed-96f3-c915e75906b0-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.897735 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ff40adb3-f509-46ed-96f3-c915e75906b0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.897866 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.897950 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.898006 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8xhm\" (UniqueName: \"kubernetes.io/projected/ff40adb3-f509-46ed-96f3-c915e75906b0-kube-api-access-n8xhm\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.898109 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-config\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.898216 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.898283 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ff40adb3-f509-46ed-96f3-c915e75906b0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.898338 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ff40adb3-f509-46ed-96f3-c915e75906b0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.898373 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.898451 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.899065 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ff40adb3-f509-46ed-96f3-c915e75906b0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.899213 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ff40adb3-f509-46ed-96f3-c915e75906b0-prometheus-metric-storage-db\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.901609 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ff40adb3-f509-46ed-96f3-c915e75906b0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.901659 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ff40adb3-f509-46ed-96f3-c915e75906b0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.901665 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.901748 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.903181 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.903694 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.911138 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-config\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.911172 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/ff40adb3-f509-46ed-96f3-c915e75906b0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:16 crc kubenswrapper[4876]: I1205 09:17:16.919103 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8xhm\" (UniqueName: \"kubernetes.io/projected/ff40adb3-f509-46ed-96f3-c915e75906b0-kube-api-access-n8xhm\") pod \"prometheus-metric-storage-0\" (UID: \"ff40adb3-f509-46ed-96f3-c915e75906b0\") " pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:17 crc kubenswrapper[4876]: I1205 09:17:17.013595 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:17 crc kubenswrapper[4876]: I1205 09:17:17.817108 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 05 09:17:17 crc kubenswrapper[4876]: I1205 09:17:17.838731 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="189edc37-9fe6-4b23-8ee4-02b72a4bb17d" path="/var/lib/kubelet/pods/189edc37-9fe6-4b23-8ee4-02b72a4bb17d/volumes" Dec 05 09:17:18 crc kubenswrapper[4876]: I1205 09:17:18.634389 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ff40adb3-f509-46ed-96f3-c915e75906b0","Type":"ContainerStarted","Data":"19ba102448fa06bfc61ec573fa4123d1566a5703957d1df9f122d63b6fb5a08b"} Dec 05 09:17:20 crc kubenswrapper[4876]: I1205 09:17:20.823566 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:17:20 crc kubenswrapper[4876]: E1205 09:17:20.824098 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:17:21 crc kubenswrapper[4876]: I1205 09:17:21.666260 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ff40adb3-f509-46ed-96f3-c915e75906b0","Type":"ContainerStarted","Data":"9a86d0109d70bc1ec172296c6154c296fe3b409edb2a003c8eb6932a9ed42915"} Dec 05 09:17:28 crc kubenswrapper[4876]: I1205 09:17:28.738328 4876 generic.go:334] "Generic (PLEG): container finished" podID="ff40adb3-f509-46ed-96f3-c915e75906b0" containerID="9a86d0109d70bc1ec172296c6154c296fe3b409edb2a003c8eb6932a9ed42915" exitCode=0 Dec 05 09:17:28 crc kubenswrapper[4876]: I1205 09:17:28.738437 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ff40adb3-f509-46ed-96f3-c915e75906b0","Type":"ContainerDied","Data":"9a86d0109d70bc1ec172296c6154c296fe3b409edb2a003c8eb6932a9ed42915"} Dec 05 09:17:29 crc kubenswrapper[4876]: I1205 09:17:29.748461 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ff40adb3-f509-46ed-96f3-c915e75906b0","Type":"ContainerStarted","Data":"9287975ab0a14ac4c528427eff52bf73590acca594705c0ea390e6dd17ba858c"} Dec 05 09:17:31 crc kubenswrapper[4876]: I1205 09:17:31.823612 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:17:31 crc kubenswrapper[4876]: E1205 09:17:31.824174 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:17:32 crc kubenswrapper[4876]: I1205 09:17:32.798806 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ff40adb3-f509-46ed-96f3-c915e75906b0","Type":"ContainerStarted","Data":"9b07def40232a5da66e8afe0c4d633f203afdf8084e101e2af6524f2b6fc6a70"} Dec 05 09:17:32 crc kubenswrapper[4876]: I1205 09:17:32.799274 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"ff40adb3-f509-46ed-96f3-c915e75906b0","Type":"ContainerStarted","Data":"e95fa08947f453db041f08760b16fa04c690307eebb73b9756468b3c4e1d2912"} Dec 05 09:17:32 crc kubenswrapper[4876]: I1205 09:17:32.837503 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.83748287 podStartE2EDuration="16.83748287s" podCreationTimestamp="2025-12-05 09:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:17:32.827053086 +0000 UTC m=+3717.315717718" watchObservedRunningTime="2025-12-05 09:17:32.83748287 +0000 UTC m=+3717.326147492" Dec 05 09:17:37 crc kubenswrapper[4876]: I1205 09:17:37.013760 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:46 crc kubenswrapper[4876]: I1205 09:17:46.824257 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:17:46 crc kubenswrapper[4876]: E1205 09:17:46.825019 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:17:47 crc kubenswrapper[4876]: I1205 09:17:47.013886 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:47 crc kubenswrapper[4876]: I1205 09:17:47.019577 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 05 09:17:47 crc kubenswrapper[4876]: I1205 09:17:47.960544 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 05 09:18:01 crc kubenswrapper[4876]: I1205 09:18:01.824279 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:18:01 crc kubenswrapper[4876]: E1205 09:18:01.825200 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:18:13 crc kubenswrapper[4876]: I1205 09:18:13.823838 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:18:13 crc kubenswrapper[4876]: E1205 09:18:13.824576 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:18:25 crc kubenswrapper[4876]: I1205 09:18:25.828254 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:18:25 crc kubenswrapper[4876]: E1205 09:18:25.829285 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:18:38 crc kubenswrapper[4876]: I1205 09:18:38.824601 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:18:38 crc kubenswrapper[4876]: E1205 09:18:38.826007 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:18:49 crc kubenswrapper[4876]: I1205 09:18:49.823875 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:18:49 crc kubenswrapper[4876]: E1205 09:18:49.824863 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:19:04 crc kubenswrapper[4876]: I1205 09:19:04.824014 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:19:04 crc kubenswrapper[4876]: E1205 09:19:04.824873 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:19:15 crc kubenswrapper[4876]: I1205 09:19:15.216934 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/manager/0.log" Dec 05 09:19:18 crc kubenswrapper[4876]: I1205 09:19:18.825142 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:19:18 crc kubenswrapper[4876]: E1205 09:19:18.828008 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:19:32 crc kubenswrapper[4876]: I1205 09:19:32.824354 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:19:32 crc kubenswrapper[4876]: E1205 09:19:32.825334 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:19:38 crc kubenswrapper[4876]: I1205 09:19:38.890613 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kv4pt/must-gather-b24tw"] Dec 05 09:19:38 crc kubenswrapper[4876]: I1205 09:19:38.892835 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:19:38 crc kubenswrapper[4876]: I1205 09:19:38.895619 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-kv4pt"/"default-dockercfg-56mtj" Dec 05 09:19:38 crc kubenswrapper[4876]: I1205 09:19:38.895724 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kv4pt"/"kube-root-ca.crt" Dec 05 09:19:38 crc kubenswrapper[4876]: I1205 09:19:38.895865 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kv4pt"/"openshift-service-ca.crt" Dec 05 09:19:38 crc kubenswrapper[4876]: I1205 09:19:38.899921 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kv4pt/must-gather-b24tw"] Dec 05 09:19:39 crc kubenswrapper[4876]: I1205 09:19:39.005399 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq2nh\" (UniqueName: \"kubernetes.io/projected/fcdc43e7-3f46-493c-963c-fa640144404e-kube-api-access-mq2nh\") pod \"must-gather-b24tw\" (UID: \"fcdc43e7-3f46-493c-963c-fa640144404e\") " pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:19:39 crc kubenswrapper[4876]: I1205 09:19:39.007125 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fcdc43e7-3f46-493c-963c-fa640144404e-must-gather-output\") pod \"must-gather-b24tw\" (UID: \"fcdc43e7-3f46-493c-963c-fa640144404e\") " pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:19:39 crc kubenswrapper[4876]: I1205 09:19:39.109460 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq2nh\" (UniqueName: \"kubernetes.io/projected/fcdc43e7-3f46-493c-963c-fa640144404e-kube-api-access-mq2nh\") pod \"must-gather-b24tw\" (UID: \"fcdc43e7-3f46-493c-963c-fa640144404e\") " pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:19:39 crc kubenswrapper[4876]: I1205 09:19:39.109623 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fcdc43e7-3f46-493c-963c-fa640144404e-must-gather-output\") pod \"must-gather-b24tw\" (UID: \"fcdc43e7-3f46-493c-963c-fa640144404e\") " pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:19:39 crc kubenswrapper[4876]: I1205 09:19:39.110133 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fcdc43e7-3f46-493c-963c-fa640144404e-must-gather-output\") pod \"must-gather-b24tw\" (UID: \"fcdc43e7-3f46-493c-963c-fa640144404e\") " pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:19:39 crc kubenswrapper[4876]: I1205 09:19:39.130337 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq2nh\" (UniqueName: \"kubernetes.io/projected/fcdc43e7-3f46-493c-963c-fa640144404e-kube-api-access-mq2nh\") pod \"must-gather-b24tw\" (UID: \"fcdc43e7-3f46-493c-963c-fa640144404e\") " pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:19:39 crc kubenswrapper[4876]: I1205 09:19:39.253273 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:19:40 crc kubenswrapper[4876]: I1205 09:19:40.162702 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kv4pt/must-gather-b24tw"] Dec 05 09:19:40 crc kubenswrapper[4876]: I1205 09:19:40.190510 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/must-gather-b24tw" event={"ID":"fcdc43e7-3f46-493c-963c-fa640144404e","Type":"ContainerStarted","Data":"615d5ce9a411499fe2dfa2e26b454356736350b5679105b31d14c893e9b8ad93"} Dec 05 09:19:45 crc kubenswrapper[4876]: I1205 09:19:45.248820 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/must-gather-b24tw" event={"ID":"fcdc43e7-3f46-493c-963c-fa640144404e","Type":"ContainerStarted","Data":"0e96b49683f81aa4b91ac7e72cdfedf421ebe4faf51aa5135aabecb7858bd3b7"} Dec 05 09:19:45 crc kubenswrapper[4876]: I1205 09:19:45.249383 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/must-gather-b24tw" event={"ID":"fcdc43e7-3f46-493c-963c-fa640144404e","Type":"ContainerStarted","Data":"452f47735e63f94a1ae2ccbc7d2c35e7f693033b26d771fc6035d34e0e9a9619"} Dec 05 09:19:45 crc kubenswrapper[4876]: I1205 09:19:45.267229 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-kv4pt/must-gather-b24tw" podStartSLOduration=3.12712929 podStartE2EDuration="7.267211818s" podCreationTimestamp="2025-12-05 09:19:38 +0000 UTC" firstStartedPulling="2025-12-05 09:19:40.158917319 +0000 UTC m=+3844.647581941" lastFinishedPulling="2025-12-05 09:19:44.298999847 +0000 UTC m=+3848.787664469" observedRunningTime="2025-12-05 09:19:45.261618138 +0000 UTC m=+3849.750282760" watchObservedRunningTime="2025-12-05 09:19:45.267211818 +0000 UTC m=+3849.755876440" Dec 05 09:19:46 crc kubenswrapper[4876]: I1205 09:19:46.824241 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:19:46 crc kubenswrapper[4876]: E1205 09:19:46.824821 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:19:48 crc kubenswrapper[4876]: E1205 09:19:48.737665 4876 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.234:55516->38.102.83.234:43593: read tcp 38.102.83.234:55516->38.102.83.234:43593: read: connection reset by peer Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.502569 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kv4pt/crc-debug-9wr9l"] Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.503832 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.627565 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bfkh\" (UniqueName: \"kubernetes.io/projected/eaa11f6b-cac4-4d14-b020-bcc69a5af077-kube-api-access-5bfkh\") pod \"crc-debug-9wr9l\" (UID: \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\") " pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.628067 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaa11f6b-cac4-4d14-b020-bcc69a5af077-host\") pod \"crc-debug-9wr9l\" (UID: \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\") " pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.729487 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bfkh\" (UniqueName: \"kubernetes.io/projected/eaa11f6b-cac4-4d14-b020-bcc69a5af077-kube-api-access-5bfkh\") pod \"crc-debug-9wr9l\" (UID: \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\") " pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.729609 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaa11f6b-cac4-4d14-b020-bcc69a5af077-host\") pod \"crc-debug-9wr9l\" (UID: \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\") " pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.729787 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaa11f6b-cac4-4d14-b020-bcc69a5af077-host\") pod \"crc-debug-9wr9l\" (UID: \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\") " pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.750929 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bfkh\" (UniqueName: \"kubernetes.io/projected/eaa11f6b-cac4-4d14-b020-bcc69a5af077-kube-api-access-5bfkh\") pod \"crc-debug-9wr9l\" (UID: \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\") " pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:19:49 crc kubenswrapper[4876]: I1205 09:19:49.826692 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:19:49 crc kubenswrapper[4876]: W1205 09:19:49.888144 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeaa11f6b_cac4_4d14_b020_bcc69a5af077.slice/crio-b726feab202a03d18b0436153bd48ff97698602d14870d3c5397b452e4c49aca WatchSource:0}: Error finding container b726feab202a03d18b0436153bd48ff97698602d14870d3c5397b452e4c49aca: Status 404 returned error can't find the container with id b726feab202a03d18b0436153bd48ff97698602d14870d3c5397b452e4c49aca Dec 05 09:19:50 crc kubenswrapper[4876]: I1205 09:19:50.306333 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" event={"ID":"eaa11f6b-cac4-4d14-b020-bcc69a5af077","Type":"ContainerStarted","Data":"b726feab202a03d18b0436153bd48ff97698602d14870d3c5397b452e4c49aca"} Dec 05 09:19:59 crc kubenswrapper[4876]: I1205 09:19:59.824428 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:19:59 crc kubenswrapper[4876]: E1205 09:19:59.825235 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:20:02 crc kubenswrapper[4876]: I1205 09:20:02.440055 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" event={"ID":"eaa11f6b-cac4-4d14-b020-bcc69a5af077","Type":"ContainerStarted","Data":"06c7e13827f955085ea031930c2ac21bd4e664534be26d6db495bbfe2059630b"} Dec 05 09:20:02 crc kubenswrapper[4876]: I1205 09:20:02.467225 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" podStartSLOduration=1.8897774360000001 podStartE2EDuration="13.467206444s" podCreationTimestamp="2025-12-05 09:19:49 +0000 UTC" firstStartedPulling="2025-12-05 09:19:49.89056928 +0000 UTC m=+3854.379233902" lastFinishedPulling="2025-12-05 09:20:01.467998288 +0000 UTC m=+3865.956662910" observedRunningTime="2025-12-05 09:20:02.454365548 +0000 UTC m=+3866.943030170" watchObservedRunningTime="2025-12-05 09:20:02.467206444 +0000 UTC m=+3866.955871066" Dec 05 09:20:13 crc kubenswrapper[4876]: I1205 09:20:13.823801 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:20:13 crc kubenswrapper[4876]: E1205 09:20:13.824654 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:20:24 crc kubenswrapper[4876]: I1205 09:20:24.859182 4876 generic.go:334] "Generic (PLEG): container finished" podID="eaa11f6b-cac4-4d14-b020-bcc69a5af077" containerID="06c7e13827f955085ea031930c2ac21bd4e664534be26d6db495bbfe2059630b" exitCode=0 Dec 05 09:20:24 crc kubenswrapper[4876]: I1205 09:20:24.859254 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" event={"ID":"eaa11f6b-cac4-4d14-b020-bcc69a5af077","Type":"ContainerDied","Data":"06c7e13827f955085ea031930c2ac21bd4e664534be26d6db495bbfe2059630b"} Dec 05 09:20:25 crc kubenswrapper[4876]: I1205 09:20:25.998018 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.055984 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kv4pt/crc-debug-9wr9l"] Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.072813 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kv4pt/crc-debug-9wr9l"] Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.136601 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaa11f6b-cac4-4d14-b020-bcc69a5af077-host\") pod \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\" (UID: \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\") " Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.136720 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaa11f6b-cac4-4d14-b020-bcc69a5af077-host" (OuterVolumeSpecName: "host") pod "eaa11f6b-cac4-4d14-b020-bcc69a5af077" (UID: "eaa11f6b-cac4-4d14-b020-bcc69a5af077"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.136754 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bfkh\" (UniqueName: \"kubernetes.io/projected/eaa11f6b-cac4-4d14-b020-bcc69a5af077-kube-api-access-5bfkh\") pod \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\" (UID: \"eaa11f6b-cac4-4d14-b020-bcc69a5af077\") " Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.137356 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaa11f6b-cac4-4d14-b020-bcc69a5af077-host\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.154118 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa11f6b-cac4-4d14-b020-bcc69a5af077-kube-api-access-5bfkh" (OuterVolumeSpecName: "kube-api-access-5bfkh") pod "eaa11f6b-cac4-4d14-b020-bcc69a5af077" (UID: "eaa11f6b-cac4-4d14-b020-bcc69a5af077"). InnerVolumeSpecName "kube-api-access-5bfkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.239495 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bfkh\" (UniqueName: \"kubernetes.io/projected/eaa11f6b-cac4-4d14-b020-bcc69a5af077-kube-api-access-5bfkh\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.876797 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b726feab202a03d18b0436153bd48ff97698602d14870d3c5397b452e4c49aca" Dec 05 09:20:26 crc kubenswrapper[4876]: I1205 09:20:26.876851 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/crc-debug-9wr9l" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.220226 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kv4pt/crc-debug-dvkxx"] Dec 05 09:20:27 crc kubenswrapper[4876]: E1205 09:20:27.220667 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa11f6b-cac4-4d14-b020-bcc69a5af077" containerName="container-00" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.220678 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa11f6b-cac4-4d14-b020-bcc69a5af077" containerName="container-00" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.220884 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa11f6b-cac4-4d14-b020-bcc69a5af077" containerName="container-00" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.221596 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.410167 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0338f666-799e-4c4a-8680-de20d669421e-host\") pod \"crc-debug-dvkxx\" (UID: \"0338f666-799e-4c4a-8680-de20d669421e\") " pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.410292 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtpck\" (UniqueName: \"kubernetes.io/projected/0338f666-799e-4c4a-8680-de20d669421e-kube-api-access-gtpck\") pod \"crc-debug-dvkxx\" (UID: \"0338f666-799e-4c4a-8680-de20d669421e\") " pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.512582 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0338f666-799e-4c4a-8680-de20d669421e-host\") pod \"crc-debug-dvkxx\" (UID: \"0338f666-799e-4c4a-8680-de20d669421e\") " pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.512649 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtpck\" (UniqueName: \"kubernetes.io/projected/0338f666-799e-4c4a-8680-de20d669421e-kube-api-access-gtpck\") pod \"crc-debug-dvkxx\" (UID: \"0338f666-799e-4c4a-8680-de20d669421e\") " pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.513126 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0338f666-799e-4c4a-8680-de20d669421e-host\") pod \"crc-debug-dvkxx\" (UID: \"0338f666-799e-4c4a-8680-de20d669421e\") " pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.541059 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtpck\" (UniqueName: \"kubernetes.io/projected/0338f666-799e-4c4a-8680-de20d669421e-kube-api-access-gtpck\") pod \"crc-debug-dvkxx\" (UID: \"0338f666-799e-4c4a-8680-de20d669421e\") " pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.834869 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaa11f6b-cac4-4d14-b020-bcc69a5af077" path="/var/lib/kubelet/pods/eaa11f6b-cac4-4d14-b020-bcc69a5af077/volumes" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.835827 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:27 crc kubenswrapper[4876]: I1205 09:20:27.889526 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" event={"ID":"0338f666-799e-4c4a-8680-de20d669421e","Type":"ContainerStarted","Data":"4333155a1d1ac7aa6144ccf1d3694d41593923c015ab908b2320e5b966e17b6e"} Dec 05 09:20:28 crc kubenswrapper[4876]: I1205 09:20:28.824882 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:20:28 crc kubenswrapper[4876]: E1205 09:20:28.825558 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:20:28 crc kubenswrapper[4876]: I1205 09:20:28.952151 4876 generic.go:334] "Generic (PLEG): container finished" podID="0338f666-799e-4c4a-8680-de20d669421e" containerID="b57ed65819ecd37b1b45bd0b3c1f0a8b58584c3769a38c5f4c67fb16b041d83e" exitCode=1 Dec 05 09:20:28 crc kubenswrapper[4876]: I1205 09:20:28.952282 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" event={"ID":"0338f666-799e-4c4a-8680-de20d669421e","Type":"ContainerDied","Data":"b57ed65819ecd37b1b45bd0b3c1f0a8b58584c3769a38c5f4c67fb16b041d83e"} Dec 05 09:20:29 crc kubenswrapper[4876]: I1205 09:20:29.000955 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kv4pt/crc-debug-dvkxx"] Dec 05 09:20:29 crc kubenswrapper[4876]: I1205 09:20:29.011272 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kv4pt/crc-debug-dvkxx"] Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.071542 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.171349 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtpck\" (UniqueName: \"kubernetes.io/projected/0338f666-799e-4c4a-8680-de20d669421e-kube-api-access-gtpck\") pod \"0338f666-799e-4c4a-8680-de20d669421e\" (UID: \"0338f666-799e-4c4a-8680-de20d669421e\") " Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.171407 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0338f666-799e-4c4a-8680-de20d669421e-host\") pod \"0338f666-799e-4c4a-8680-de20d669421e\" (UID: \"0338f666-799e-4c4a-8680-de20d669421e\") " Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.171826 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0338f666-799e-4c4a-8680-de20d669421e-host" (OuterVolumeSpecName: "host") pod "0338f666-799e-4c4a-8680-de20d669421e" (UID: "0338f666-799e-4c4a-8680-de20d669421e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.179269 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0338f666-799e-4c4a-8680-de20d669421e-kube-api-access-gtpck" (OuterVolumeSpecName: "kube-api-access-gtpck") pod "0338f666-799e-4c4a-8680-de20d669421e" (UID: "0338f666-799e-4c4a-8680-de20d669421e"). InnerVolumeSpecName "kube-api-access-gtpck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.273443 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtpck\" (UniqueName: \"kubernetes.io/projected/0338f666-799e-4c4a-8680-de20d669421e-kube-api-access-gtpck\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.273482 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0338f666-799e-4c4a-8680-de20d669421e-host\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.971661 4876 scope.go:117] "RemoveContainer" containerID="b57ed65819ecd37b1b45bd0b3c1f0a8b58584c3769a38c5f4c67fb16b041d83e" Dec 05 09:20:30 crc kubenswrapper[4876]: I1205 09:20:30.972026 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/crc-debug-dvkxx" Dec 05 09:20:31 crc kubenswrapper[4876]: I1205 09:20:31.838391 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0338f666-799e-4c4a-8680-de20d669421e" path="/var/lib/kubelet/pods/0338f666-799e-4c4a-8680-de20d669421e/volumes" Dec 05 09:20:39 crc kubenswrapper[4876]: I1205 09:20:39.824443 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:20:40 crc kubenswrapper[4876]: I1205 09:20:40.065382 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"e6e7b471536bd5685c4904d6dc57a671c171f7553fd3d753fe9eabbc7e548d4f"} Dec 05 09:21:14 crc kubenswrapper[4876]: I1205 09:21:14.496933 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a9faf08a-6b0e-435c-9810-d94212aa82ad/init-config-reloader/0.log" Dec 05 09:21:14 crc kubenswrapper[4876]: I1205 09:21:14.670762 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a9faf08a-6b0e-435c-9810-d94212aa82ad/init-config-reloader/0.log" Dec 05 09:21:14 crc kubenswrapper[4876]: I1205 09:21:14.729834 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a9faf08a-6b0e-435c-9810-d94212aa82ad/alertmanager/0.log" Dec 05 09:21:14 crc kubenswrapper[4876]: I1205 09:21:14.760150 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a9faf08a-6b0e-435c-9810-d94212aa82ad/config-reloader/0.log" Dec 05 09:21:14 crc kubenswrapper[4876]: I1205 09:21:14.931427 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e960eaad-8be1-47d2-8e07-a17915f9cfea/aodh-api/0.log" Dec 05 09:21:14 crc kubenswrapper[4876]: I1205 09:21:14.972490 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e960eaad-8be1-47d2-8e07-a17915f9cfea/aodh-evaluator/0.log" Dec 05 09:21:14 crc kubenswrapper[4876]: I1205 09:21:14.995590 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e960eaad-8be1-47d2-8e07-a17915f9cfea/aodh-listener/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.081639 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e960eaad-8be1-47d2-8e07-a17915f9cfea/aodh-notifier/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.170291 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9d7c76978-sl6zr_d9c4abfe-e5d6-4a04-88ff-3e721675a75e/barbican-api/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.233354 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9d7c76978-sl6zr_d9c4abfe-e5d6-4a04-88ff-3e721675a75e/barbican-api-log/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.398519 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-69575954c8-dwfz7_23f75a46-ea18-4217-bcb8-26d5652f95eb/barbican-keystone-listener/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.475597 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-69575954c8-dwfz7_23f75a46-ea18-4217-bcb8-26d5652f95eb/barbican-keystone-listener-log/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.574967 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f66c6dbdf-59mzz_aa8ee581-b9fc-4c2d-8139-b3d827f71697/barbican-worker/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.656161 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f66c6dbdf-59mzz_aa8ee581-b9fc-4c2d-8139-b3d827f71697/barbican-worker-log/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.740591 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj_5f7a4eab-9eaf-4599-8eb6-deef38e971c9/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.903077 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2143f266-92fd-4c74-b305-0e1d95d3044b/ceilometer-central-agent/0.log" Dec 05 09:21:15 crc kubenswrapper[4876]: I1205 09:21:15.981969 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2143f266-92fd-4c74-b305-0e1d95d3044b/ceilometer-notification-agent/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.014004 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2143f266-92fd-4c74-b305-0e1d95d3044b/proxy-httpd/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.130061 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2143f266-92fd-4c74-b305-0e1d95d3044b/sg-core/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.255990 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9e321464-50bf-42c0-8584-a75e9282eb65/cinder-api-log/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.297449 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9e321464-50bf-42c0-8584-a75e9282eb65/cinder-api/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.494086 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ba584dcd-132e-407b-960e-4a68cfb1ea31/probe/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.496986 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ba584dcd-132e-407b-960e-4a68cfb1ea31/cinder-scheduler/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.719411 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-pfqln_9d51e73b-805b-4690-aed5-514789933cce/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.742736 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z_8a524784-c668-4388-81c5-719f97dbb947/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:16 crc kubenswrapper[4876]: I1205 09:21:16.966550 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-dcb46_ece6a347-25cd-41e9-89e0-f20c1fdb9eba/init/0.log" Dec 05 09:21:17 crc kubenswrapper[4876]: I1205 09:21:17.146533 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-dcb46_ece6a347-25cd-41e9-89e0-f20c1fdb9eba/init/0.log" Dec 05 09:21:17 crc kubenswrapper[4876]: I1205 09:21:17.162202 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-dcb46_ece6a347-25cd-41e9-89e0-f20c1fdb9eba/dnsmasq-dns/0.log" Dec 05 09:21:17 crc kubenswrapper[4876]: I1205 09:21:17.197552 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn_d17d4141-7f2f-4598-9cf4-96870bca5903/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:17 crc kubenswrapper[4876]: I1205 09:21:17.353634 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e70a9dae-3695-42fa-bab0-b0ef8fc44e13/glance-httpd/0.log" Dec 05 09:21:17 crc kubenswrapper[4876]: I1205 09:21:17.410060 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e70a9dae-3695-42fa-bab0-b0ef8fc44e13/glance-log/0.log" Dec 05 09:21:17 crc kubenswrapper[4876]: I1205 09:21:17.755589 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_00302ae2-a0ed-4a18-80e4-67ac2fed0230/glance-httpd/0.log" Dec 05 09:21:17 crc kubenswrapper[4876]: I1205 09:21:17.836669 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_00302ae2-a0ed-4a18-80e4-67ac2fed0230/glance-log/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.240412 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5f87bc649b-spcsf_b7673120-d39b-416e-8f5f-260cdce6f71b/heat-api/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.371732 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-b64b8b465-trj6k_6a133b19-0c3f-4a85-95ea-910b9c0c4df0/heat-engine/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.429106 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-76699577f6-nt9wb_0f3a1341-ca1e-4b8c-a57d-0c83d39a403e/heat-cfnapi/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.433242 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b_ecc3ee00-cb0c-4084-ba46-80684e18f07e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.704705 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-slf72_f596bae7-c567-469e-b7be-f95265c0f016/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.744297 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-577c476758-ksnrz_5612eb81-67eb-4559-b21a-549e4becca72/keystone-api/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.752910 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29415421-28x2l_778ede62-2e7e-4967-a96a-a8073b375f18/keystone-cron/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.896528 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1c06bc09-5dd5-4d2c-b572-078ea2a4d037/kube-state-metrics/0.log" Dec 05 09:21:18 crc kubenswrapper[4876]: I1205 09:21:18.995673 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl_21e86dc4-3382-464d-aa9b-a811ee65e5d1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.230318 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5ff8f8c8c7-rx6w7_cd9bce2b-4ecb-45b0-a1b2-309c3f450a47/neutron-api/0.log" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.319025 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5ff8f8c8c7-rx6w7_cd9bce2b-4ecb-45b0-a1b2-309c3f450a47/neutron-httpd/0.log" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.397952 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c_74b443c1-b03a-42d2-8a66-49510fdfb5bb/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.520540 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k469q"] Dec 05 09:21:19 crc kubenswrapper[4876]: E1205 09:21:19.521341 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0338f666-799e-4c4a-8680-de20d669421e" containerName="container-00" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.521364 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="0338f666-799e-4c4a-8680-de20d669421e" containerName="container-00" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.521579 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="0338f666-799e-4c4a-8680-de20d669421e" containerName="container-00" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.525613 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.540145 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k469q"] Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.687887 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-utilities\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.688166 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v9j9\" (UniqueName: \"kubernetes.io/projected/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-kube-api-access-8v9j9\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.688373 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-catalog-content\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.696928 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b8b58af0-0308-45aa-825c-0d252a0bff0d/nova-api-log/0.log" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.790015 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-catalog-content\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.790157 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-utilities\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.790266 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v9j9\" (UniqueName: \"kubernetes.io/projected/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-kube-api-access-8v9j9\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.790984 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-catalog-content\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.791677 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-utilities\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.817006 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v9j9\" (UniqueName: \"kubernetes.io/projected/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-kube-api-access-8v9j9\") pod \"redhat-operators-k469q\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.854239 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:19 crc kubenswrapper[4876]: I1205 09:21:19.948445 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_dc5cc0db-4ac5-4093-a0f0-abf202104b4f/nova-cell0-conductor-conductor/0.log" Dec 05 09:21:20 crc kubenswrapper[4876]: I1205 09:21:20.132783 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b0643cc1-c10a-4da7-a3b8-09063ec41b80/nova-cell1-conductor-conductor/0.log" Dec 05 09:21:20 crc kubenswrapper[4876]: I1205 09:21:20.157871 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b8b58af0-0308-45aa-825c-0d252a0bff0d/nova-api-api/0.log" Dec 05 09:21:20 crc kubenswrapper[4876]: I1205 09:21:20.396768 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k469q"] Dec 05 09:21:20 crc kubenswrapper[4876]: I1205 09:21:20.461301 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8ee1cf7e-9e7b-407a-ab48-d4abb9326b27/nova-cell1-novncproxy-novncproxy/0.log" Dec 05 09:21:20 crc kubenswrapper[4876]: I1205 09:21:20.485551 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k469q" event={"ID":"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c","Type":"ContainerStarted","Data":"856bf7f1ca16f16e2fad5f39ddef3bae6670f0279265e8b9230a9cefedd3e98e"} Dec 05 09:21:20 crc kubenswrapper[4876]: I1205 09:21:20.577872 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5hxqz_d5afc65a-8adf-4aa6-a348-469c81b51bad/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:20 crc kubenswrapper[4876]: I1205 09:21:20.922920 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7ae3be31-2c00-4486-9c64-10c7fcb2c787/nova-metadata-log/0.log" Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.236184 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_761374dd-700e-42fa-81d4-d560f97643be/mysql-bootstrap/0.log" Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.309060 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_304bb281-a6d7-460a-a81e-3cecbb60e20f/nova-scheduler-scheduler/0.log" Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.423943 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_761374dd-700e-42fa-81d4-d560f97643be/mysql-bootstrap/0.log" Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.506819 4876 generic.go:334] "Generic (PLEG): container finished" podID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerID="e3bdb6619e6883af1ec27ba33378979c71db2abe8c8828581396119183923322" exitCode=0 Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.507326 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k469q" event={"ID":"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c","Type":"ContainerDied","Data":"e3bdb6619e6883af1ec27ba33378979c71db2abe8c8828581396119183923322"} Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.512609 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.541410 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_761374dd-700e-42fa-81d4-d560f97643be/galera/0.log" Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.656634 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d47eff06-2977-4bcb-bc3b-6ad08399bbd7/mysql-bootstrap/0.log" Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.843139 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d47eff06-2977-4bcb-bc3b-6ad08399bbd7/mysql-bootstrap/0.log" Dec 05 09:21:21 crc kubenswrapper[4876]: I1205 09:21:21.858939 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d47eff06-2977-4bcb-bc3b-6ad08399bbd7/galera/0.log" Dec 05 09:21:22 crc kubenswrapper[4876]: I1205 09:21:22.158661 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7ae3be31-2c00-4486-9c64-10c7fcb2c787/nova-metadata-metadata/0.log" Dec 05 09:21:22 crc kubenswrapper[4876]: I1205 09:21:22.238078 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-7p5wk_442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2/ovn-controller/0.log" Dec 05 09:21:22 crc kubenswrapper[4876]: I1205 09:21:22.406529 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_17ad89b3-21e5-4db2-8d47-29248aa9e054/openstackclient/0.log" Dec 05 09:21:22 crc kubenswrapper[4876]: I1205 09:21:22.562756 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hxnnn_82469e38-4872-4d09-8a1e-2e635df62905/openstack-network-exporter/0.log" Dec 05 09:21:22 crc kubenswrapper[4876]: I1205 09:21:22.744234 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhwc7_07839a94-e94b-4ff2-9eaf-9fb315c83046/ovsdb-server-init/0.log" Dec 05 09:21:22 crc kubenswrapper[4876]: I1205 09:21:22.948583 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhwc7_07839a94-e94b-4ff2-9eaf-9fb315c83046/ovsdb-server/0.log" Dec 05 09:21:22 crc kubenswrapper[4876]: I1205 09:21:22.968706 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhwc7_07839a94-e94b-4ff2-9eaf-9fb315c83046/ovsdb-server-init/0.log" Dec 05 09:21:22 crc kubenswrapper[4876]: I1205 09:21:22.993273 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhwc7_07839a94-e94b-4ff2-9eaf-9fb315c83046/ovs-vswitchd/0.log" Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.198288 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-pdx5k_33523fba-8378-49ee-b023-0f80a4a8f91b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.209510 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f6758b3-0cba-4d9b-b3ad-25149d206ff9/openstack-network-exporter/0.log" Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.308421 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f6758b3-0cba-4d9b-b3ad-25149d206ff9/ovn-northd/0.log" Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.448186 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3f18961c-c654-4f94-98a3-697644fdfda4/openstack-network-exporter/0.log" Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.449357 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3f18961c-c654-4f94-98a3-697644fdfda4/ovsdbserver-nb/0.log" Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.560862 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k469q" event={"ID":"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c","Type":"ContainerStarted","Data":"967082afac87fa43dec0b3297d2c3746207fa771135526cd2b29ed1511be6beb"} Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.742469 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9c3699c2-facd-4d82-92ec-4816b871833b/ovsdbserver-sb/0.log" Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.746718 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9c3699c2-facd-4d82-92ec-4816b871833b/openstack-network-exporter/0.log" Dec 05 09:21:23 crc kubenswrapper[4876]: I1205 09:21:23.900691 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-56d57865bd-mfljr_27487f6b-feb8-47c1-98db-5ed28120b3f3/placement-api/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.134631 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-56d57865bd-mfljr_27487f6b-feb8-47c1-98db-5ed28120b3f3/placement-log/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.146729 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/init-config-reloader/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.386915 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/config-reloader/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.387625 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/prometheus/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.433343 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/thanos-sidecar/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.469339 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/init-config-reloader/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.580964 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ef895f6-3d13-4980-98ef-9111438fe1ec/setup-container/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.909686 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ef895f6-3d13-4980-98ef-9111438fe1ec/setup-container/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.915749 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ef895f6-3d13-4980-98ef-9111438fe1ec/rabbitmq/0.log" Dec 05 09:21:24 crc kubenswrapper[4876]: I1205 09:21:24.954421 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_692469d9-b150-4d67-b412-ba8760598f8b/setup-container/0.log" Dec 05 09:21:25 crc kubenswrapper[4876]: I1205 09:21:25.169832 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_692469d9-b150-4d67-b412-ba8760598f8b/setup-container/0.log" Dec 05 09:21:25 crc kubenswrapper[4876]: I1205 09:21:25.205344 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w_a4528785-cb5d-4281-9de7-2140359d7c0c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:25 crc kubenswrapper[4876]: I1205 09:21:25.205670 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_692469d9-b150-4d67-b412-ba8760598f8b/rabbitmq/0.log" Dec 05 09:21:25 crc kubenswrapper[4876]: I1205 09:21:25.520197 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-7xhsh_fde6a633-2ee9-4fad-b157-24435a2e89d4/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:25 crc kubenswrapper[4876]: I1205 09:21:25.539133 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz_4554e5a3-3a60-4bc6-85c7-bfdc19421f05/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:25 crc kubenswrapper[4876]: I1205 09:21:25.955622 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-stmmf_dc2909c6-f489-467c-a203-01adb27428f5/ssh-known-hosts-edpm-deployment/0.log" Dec 05 09:21:25 crc kubenswrapper[4876]: I1205 09:21:25.961143 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-8w987_87e85de2-b888-48c3-8f2d-e500ccedca3e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.301644 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-98b56b8f5-ksg4z_fbd7f2e5-b830-45be-be0b-1f23806634e2/proxy-server/0.log" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.307762 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jdckq"] Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.323686 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.324321 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-98b56b8f5-ksg4z_fbd7f2e5-b830-45be-be0b-1f23806634e2/proxy-httpd/0.log" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.343358 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jdckq"] Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.437690 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4swlc\" (UniqueName: \"kubernetes.io/projected/b09de4f9-07c4-41d6-801c-872380203cc5-kube-api-access-4swlc\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.437756 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-utilities\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.437790 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-catalog-content\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.474775 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-hzjsf_d41c7aad-b564-4c13-9f60-2d1fe53861a8/swift-ring-rebalance/0.log" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.541547 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-catalog-content\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.541794 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4swlc\" (UniqueName: \"kubernetes.io/projected/b09de4f9-07c4-41d6-801c-872380203cc5-kube-api-access-4swlc\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.542028 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-utilities\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.542564 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-catalog-content\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.542633 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-utilities\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.589982 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4swlc\" (UniqueName: \"kubernetes.io/projected/b09de4f9-07c4-41d6-801c-872380203cc5-kube-api-access-4swlc\") pod \"community-operators-jdckq\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.641320 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/account-reaper/0.log" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.657638 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/account-auditor/0.log" Dec 05 09:21:26 crc kubenswrapper[4876]: I1205 09:21:26.662331 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.034287 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/container-auditor/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.063448 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/account-server/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.135563 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/account-replicator/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.206189 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/container-replicator/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.252465 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/container-server/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.289478 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jdckq"] Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.325396 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/container-updater/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.432240 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-auditor/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.486793 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-replicator/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.546971 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-expirer/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.619201 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-server/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.646118 4876 generic.go:334] "Generic (PLEG): container finished" podID="b09de4f9-07c4-41d6-801c-872380203cc5" containerID="fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad" exitCode=0 Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.646187 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdckq" event={"ID":"b09de4f9-07c4-41d6-801c-872380203cc5","Type":"ContainerDied","Data":"fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad"} Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.646213 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdckq" event={"ID":"b09de4f9-07c4-41d6-801c-872380203cc5","Type":"ContainerStarted","Data":"f3abcee9dbde60ea77728e2a21b99e07ae1cac9c45fd5a3c54f8a09e0048c4e3"} Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.664625 4876 generic.go:334] "Generic (PLEG): container finished" podID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerID="967082afac87fa43dec0b3297d2c3746207fa771135526cd2b29ed1511be6beb" exitCode=0 Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.664668 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k469q" event={"ID":"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c","Type":"ContainerDied","Data":"967082afac87fa43dec0b3297d2c3746207fa771135526cd2b29ed1511be6beb"} Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.697640 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/rsync/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.700295 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-updater/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.798333 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/swift-recon-cron/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.994887 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-zc47n_611e9ca4-b50e-415e-bf76-ec70f386af72/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:27 crc kubenswrapper[4876]: I1205 09:21:27.999733 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-77bdc_91a239cd-2756-4d03-8175-32dd507c0b2a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:21:28 crc kubenswrapper[4876]: I1205 09:21:28.675631 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k469q" event={"ID":"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c","Type":"ContainerStarted","Data":"e113f5cf129e4314f07f57a05f0cf3509efd5a331aff56037b1073fc0bbd323c"} Dec 05 09:21:28 crc kubenswrapper[4876]: I1205 09:21:28.678506 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdckq" event={"ID":"b09de4f9-07c4-41d6-801c-872380203cc5","Type":"ContainerStarted","Data":"dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7"} Dec 05 09:21:28 crc kubenswrapper[4876]: I1205 09:21:28.709370 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k469q" podStartSLOduration=3.100291048 podStartE2EDuration="9.709349517s" podCreationTimestamp="2025-12-05 09:21:19 +0000 UTC" firstStartedPulling="2025-12-05 09:21:21.512353346 +0000 UTC m=+3946.001017968" lastFinishedPulling="2025-12-05 09:21:28.121411815 +0000 UTC m=+3952.610076437" observedRunningTime="2025-12-05 09:21:28.698512945 +0000 UTC m=+3953.187177567" watchObservedRunningTime="2025-12-05 09:21:28.709349517 +0000 UTC m=+3953.198014159" Dec 05 09:21:29 crc kubenswrapper[4876]: I1205 09:21:29.693959 4876 generic.go:334] "Generic (PLEG): container finished" podID="b09de4f9-07c4-41d6-801c-872380203cc5" containerID="dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7" exitCode=0 Dec 05 09:21:29 crc kubenswrapper[4876]: I1205 09:21:29.694136 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdckq" event={"ID":"b09de4f9-07c4-41d6-801c-872380203cc5","Type":"ContainerDied","Data":"dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7"} Dec 05 09:21:29 crc kubenswrapper[4876]: I1205 09:21:29.855759 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:29 crc kubenswrapper[4876]: I1205 09:21:29.856827 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:30 crc kubenswrapper[4876]: I1205 09:21:30.712549 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdckq" event={"ID":"b09de4f9-07c4-41d6-801c-872380203cc5","Type":"ContainerStarted","Data":"96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607"} Dec 05 09:21:30 crc kubenswrapper[4876]: I1205 09:21:30.732197 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jdckq" podStartSLOduration=2.287473446 podStartE2EDuration="4.732179504s" podCreationTimestamp="2025-12-05 09:21:26 +0000 UTC" firstStartedPulling="2025-12-05 09:21:27.647879475 +0000 UTC m=+3952.136544097" lastFinishedPulling="2025-12-05 09:21:30.092585533 +0000 UTC m=+3954.581250155" observedRunningTime="2025-12-05 09:21:30.728043433 +0000 UTC m=+3955.216708045" watchObservedRunningTime="2025-12-05 09:21:30.732179504 +0000 UTC m=+3955.220844126" Dec 05 09:21:30 crc kubenswrapper[4876]: I1205 09:21:30.917195 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k469q" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="registry-server" probeResult="failure" output=< Dec 05 09:21:30 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 09:21:30 crc kubenswrapper[4876]: > Dec 05 09:21:36 crc kubenswrapper[4876]: I1205 09:21:36.662529 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:36 crc kubenswrapper[4876]: I1205 09:21:36.663013 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:36 crc kubenswrapper[4876]: I1205 09:21:36.717071 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:36 crc kubenswrapper[4876]: I1205 09:21:36.837119 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:36 crc kubenswrapper[4876]: I1205 09:21:36.953812 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jdckq"] Dec 05 09:21:38 crc kubenswrapper[4876]: I1205 09:21:38.459596 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_846c521d-f570-4731-a923-8a1e9c626659/memcached/0.log" Dec 05 09:21:38 crc kubenswrapper[4876]: I1205 09:21:38.803931 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jdckq" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" containerName="registry-server" containerID="cri-o://96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607" gracePeriod=2 Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.335556 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.514011 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-utilities\") pod \"b09de4f9-07c4-41d6-801c-872380203cc5\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.514109 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4swlc\" (UniqueName: \"kubernetes.io/projected/b09de4f9-07c4-41d6-801c-872380203cc5-kube-api-access-4swlc\") pod \"b09de4f9-07c4-41d6-801c-872380203cc5\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.514180 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-catalog-content\") pod \"b09de4f9-07c4-41d6-801c-872380203cc5\" (UID: \"b09de4f9-07c4-41d6-801c-872380203cc5\") " Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.514650 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-utilities" (OuterVolumeSpecName: "utilities") pod "b09de4f9-07c4-41d6-801c-872380203cc5" (UID: "b09de4f9-07c4-41d6-801c-872380203cc5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.523169 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b09de4f9-07c4-41d6-801c-872380203cc5-kube-api-access-4swlc" (OuterVolumeSpecName: "kube-api-access-4swlc") pod "b09de4f9-07c4-41d6-801c-872380203cc5" (UID: "b09de4f9-07c4-41d6-801c-872380203cc5"). InnerVolumeSpecName "kube-api-access-4swlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.562108 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b09de4f9-07c4-41d6-801c-872380203cc5" (UID: "b09de4f9-07c4-41d6-801c-872380203cc5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.616079 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.616131 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4swlc\" (UniqueName: \"kubernetes.io/projected/b09de4f9-07c4-41d6-801c-872380203cc5-kube-api-access-4swlc\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.616144 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b09de4f9-07c4-41d6-801c-872380203cc5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.815053 4876 generic.go:334] "Generic (PLEG): container finished" podID="b09de4f9-07c4-41d6-801c-872380203cc5" containerID="96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607" exitCode=0 Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.815111 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdckq" event={"ID":"b09de4f9-07c4-41d6-801c-872380203cc5","Type":"ContainerDied","Data":"96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607"} Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.815130 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdckq" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.815153 4876 scope.go:117] "RemoveContainer" containerID="96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.815143 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdckq" event={"ID":"b09de4f9-07c4-41d6-801c-872380203cc5","Type":"ContainerDied","Data":"f3abcee9dbde60ea77728e2a21b99e07ae1cac9c45fd5a3c54f8a09e0048c4e3"} Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.867889 4876 scope.go:117] "RemoveContainer" containerID="dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.883961 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jdckq"] Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.891987 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jdckq"] Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.897067 4876 scope.go:117] "RemoveContainer" containerID="fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.922248 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.952095 4876 scope.go:117] "RemoveContainer" containerID="96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607" Dec 05 09:21:39 crc kubenswrapper[4876]: E1205 09:21:39.961062 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607\": container with ID starting with 96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607 not found: ID does not exist" containerID="96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.961111 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607"} err="failed to get container status \"96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607\": rpc error: code = NotFound desc = could not find container \"96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607\": container with ID starting with 96e6575fe7c3c1cace5da7bbd434d43b714ac7ab44f9c690b0c08778a72bc607 not found: ID does not exist" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.961146 4876 scope.go:117] "RemoveContainer" containerID="dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7" Dec 05 09:21:39 crc kubenswrapper[4876]: E1205 09:21:39.961972 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7\": container with ID starting with dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7 not found: ID does not exist" containerID="dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.961995 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7"} err="failed to get container status \"dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7\": rpc error: code = NotFound desc = could not find container \"dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7\": container with ID starting with dd87a5cd0cc74ecf9e88ca1de365c23597a6e47bb0bcf1b1b302f579c33113b7 not found: ID does not exist" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.962011 4876 scope.go:117] "RemoveContainer" containerID="fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad" Dec 05 09:21:39 crc kubenswrapper[4876]: E1205 09:21:39.962281 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad\": container with ID starting with fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad not found: ID does not exist" containerID="fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.962323 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad"} err="failed to get container status \"fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad\": rpc error: code = NotFound desc = could not find container \"fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad\": container with ID starting with fc5378b0dfd137c68ea09ae4a76b6b54106fd110c02f7bdb4af02b217961faad not found: ID does not exist" Dec 05 09:21:39 crc kubenswrapper[4876]: I1205 09:21:39.978452 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:41 crc kubenswrapper[4876]: I1205 09:21:41.834651 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" path="/var/lib/kubelet/pods/b09de4f9-07c4-41d6-801c-872380203cc5/volumes" Dec 05 09:21:42 crc kubenswrapper[4876]: I1205 09:21:42.362401 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k469q"] Dec 05 09:21:42 crc kubenswrapper[4876]: I1205 09:21:42.362673 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k469q" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="registry-server" containerID="cri-o://e113f5cf129e4314f07f57a05f0cf3509efd5a331aff56037b1073fc0bbd323c" gracePeriod=2 Dec 05 09:21:42 crc kubenswrapper[4876]: I1205 09:21:42.847373 4876 generic.go:334] "Generic (PLEG): container finished" podID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerID="e113f5cf129e4314f07f57a05f0cf3509efd5a331aff56037b1073fc0bbd323c" exitCode=0 Dec 05 09:21:42 crc kubenswrapper[4876]: I1205 09:21:42.847459 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k469q" event={"ID":"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c","Type":"ContainerDied","Data":"e113f5cf129e4314f07f57a05f0cf3509efd5a331aff56037b1073fc0bbd323c"} Dec 05 09:21:42 crc kubenswrapper[4876]: I1205 09:21:42.847644 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k469q" event={"ID":"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c","Type":"ContainerDied","Data":"856bf7f1ca16f16e2fad5f39ddef3bae6670f0279265e8b9230a9cefedd3e98e"} Dec 05 09:21:42 crc kubenswrapper[4876]: I1205 09:21:42.847659 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="856bf7f1ca16f16e2fad5f39ddef3bae6670f0279265e8b9230a9cefedd3e98e" Dec 05 09:21:42 crc kubenswrapper[4876]: I1205 09:21:42.884537 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.051700 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-utilities\") pod \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.051768 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-catalog-content\") pod \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.051810 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v9j9\" (UniqueName: \"kubernetes.io/projected/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-kube-api-access-8v9j9\") pod \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\" (UID: \"00ee21fe-67b7-42fc-84ee-cbe6ce5b500c\") " Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.052706 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-utilities" (OuterVolumeSpecName: "utilities") pod "00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" (UID: "00ee21fe-67b7-42fc-84ee-cbe6ce5b500c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.057465 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-kube-api-access-8v9j9" (OuterVolumeSpecName: "kube-api-access-8v9j9") pod "00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" (UID: "00ee21fe-67b7-42fc-84ee-cbe6ce5b500c"). InnerVolumeSpecName "kube-api-access-8v9j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.153416 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.153455 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v9j9\" (UniqueName: \"kubernetes.io/projected/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-kube-api-access-8v9j9\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.169552 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" (UID: "00ee21fe-67b7-42fc-84ee-cbe6ce5b500c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.255292 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.855985 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k469q" Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.885971 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k469q"] Dec 05 09:21:43 crc kubenswrapper[4876]: I1205 09:21:43.895531 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k469q"] Dec 05 09:21:45 crc kubenswrapper[4876]: I1205 09:21:45.841701 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" path="/var/lib/kubelet/pods/00ee21fe-67b7-42fc-84ee-cbe6ce5b500c/volumes" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.033877 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/util/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.218745 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/util/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.239141 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/pull/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.255701 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/pull/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.428443 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/util/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.449276 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/pull/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.450239 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/extract/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.630430 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ppsdr_04e3edd8-f29a-48d1-84f4-81985747740f/kube-rbac-proxy/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.672177 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-d4l6m_0036dfa3-4545-43f5-a6f6-c329485b678b/kube-rbac-proxy/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.707880 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ppsdr_04e3edd8-f29a-48d1-84f4-81985747740f/manager/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.895813 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-d4l6m_0036dfa3-4545-43f5-a6f6-c329485b678b/manager/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.904672 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-jwn22_d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca/kube-rbac-proxy/0.log" Dec 05 09:21:56 crc kubenswrapper[4876]: I1205 09:21:56.974584 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-jwn22_d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca/manager/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.122262 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-bv2mb_1c4ffc06-424a-4c94-be0d-2ff27bf8d920/kube-rbac-proxy/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.236271 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-bv2mb_1c4ffc06-424a-4c94-be0d-2ff27bf8d920/manager/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.309068 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-q7r4r_291c6269-dc8a-4193-a602-2a16202db817/kube-rbac-proxy/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.403724 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-q7r4r_291c6269-dc8a-4193-a602-2a16202db817/manager/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.447688 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-vjh2v_0b8ca916-a8c5-4866-8333-ebbc0f861161/kube-rbac-proxy/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.520934 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-vjh2v_0b8ca916-a8c5-4866-8333-ebbc0f861161/manager/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.617654 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-l9jr7_679c172a-10ff-4a0e-8932-bcc1f78393d7/kube-rbac-proxy/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.843071 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-l9jr7_679c172a-10ff-4a0e-8932-bcc1f78393d7/manager/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.862436 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-qgxmr_ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7/kube-rbac-proxy/0.log" Dec 05 09:21:57 crc kubenswrapper[4876]: I1205 09:21:57.905082 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-qgxmr_ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7/manager/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.034872 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4ckhg_e21b6aa8-f703-42e0-8251-06444e6423fe/kube-rbac-proxy/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.179852 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4ckhg_e21b6aa8-f703-42e0-8251-06444e6423fe/manager/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.236814 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-v2p7b_9f436ce1-a9d9-411a-824a-85afc3ee2069/kube-rbac-proxy/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.271955 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-v2p7b_9f436ce1-a9d9-411a-824a-85afc3ee2069/manager/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.350445 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vp2x4_13b56ee6-b8fa-40bf-add8-1778f4acf30d/kube-rbac-proxy/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.445827 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vp2x4_13b56ee6-b8fa-40bf-add8-1778f4acf30d/manager/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.587465 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lxwt9_7b0b34b8-da13-4978-9a68-606135b29834/kube-rbac-proxy/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.606654 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lxwt9_7b0b34b8-da13-4978-9a68-606135b29834/manager/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.686382 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2spcs_68006b91-3398-4ff1-8c6e-ede02835c0ef/kube-rbac-proxy/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.929655 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2spcs_68006b91-3398-4ff1-8c6e-ede02835c0ef/manager/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.962425 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-svc6v_fb571bcf-872c-42f5-9a65-18c38291b062/manager/0.log" Dec 05 09:21:58 crc kubenswrapper[4876]: I1205 09:21:58.967671 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-svc6v_fb571bcf-872c-42f5-9a65-18c38291b062/kube-rbac-proxy/0.log" Dec 05 09:21:59 crc kubenswrapper[4876]: I1205 09:21:59.156086 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6_e3a38502-6fad-4383-81e8-591318a4d4d3/kube-rbac-proxy/0.log" Dec 05 09:21:59 crc kubenswrapper[4876]: I1205 09:21:59.169406 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6_e3a38502-6fad-4383-81e8-591318a4d4d3/manager/0.log" Dec 05 09:21:59 crc kubenswrapper[4876]: I1205 09:21:59.601737 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cc6d987f6-whr8k_1c603b75-2c70-4ac9-ba31-7f2f69785482/operator/0.log" Dec 05 09:21:59 crc kubenswrapper[4876]: I1205 09:21:59.702162 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qjjkm_437c91ca-0472-481b-b58f-40269f4d2c18/registry-server/0.log" Dec 05 09:21:59 crc kubenswrapper[4876]: I1205 09:21:59.906719 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lfncq_217cf4c3-b49c-4bc6-a1cd-a12589264d11/kube-rbac-proxy/0.log" Dec 05 09:22:00 crc kubenswrapper[4876]: I1205 09:22:00.061861 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lfncq_217cf4c3-b49c-4bc6-a1cd-a12589264d11/manager/0.log" Dec 05 09:22:00 crc kubenswrapper[4876]: I1205 09:22:00.204429 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wptqj_ca537166-1c8e-4616-ba9f-b97382f11d7f/kube-rbac-proxy/0.log" Dec 05 09:22:00 crc kubenswrapper[4876]: I1205 09:22:00.375796 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wptqj_ca537166-1c8e-4616-ba9f-b97382f11d7f/manager/0.log" Dec 05 09:22:00 crc kubenswrapper[4876]: I1205 09:22:00.389615 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9ptpl_af4995ab-bc40-4446-bc2e-d421fe03423c/operator/0.log" Dec 05 09:22:00 crc kubenswrapper[4876]: I1205 09:22:00.665784 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qltmw_861cead7-82be-495f-b144-7120ee08ae2d/manager/0.log" Dec 05 09:22:00 crc kubenswrapper[4876]: I1205 09:22:00.749526 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/kube-rbac-proxy/0.log" Dec 05 09:22:00 crc kubenswrapper[4876]: I1205 09:22:00.755948 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qltmw_861cead7-82be-495f-b144-7120ee08ae2d/kube-rbac-proxy/0.log" Dec 05 09:22:00 crc kubenswrapper[4876]: I1205 09:22:00.968701 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8579f747db-9t4h4_48cca8bf-c0e1-4957-885a-4cf42c1c33fe/manager/0.log" Dec 05 09:22:01 crc kubenswrapper[4876]: I1205 09:22:01.069792 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tn7mh_af99aad7-cbab-4990-8ed5-3a66c9f4b432/kube-rbac-proxy/0.log" Dec 05 09:22:01 crc kubenswrapper[4876]: I1205 09:22:01.126360 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tn7mh_af99aad7-cbab-4990-8ed5-3a66c9f4b432/manager/0.log" Dec 05 09:22:01 crc kubenswrapper[4876]: I1205 09:22:01.159630 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/manager/0.log" Dec 05 09:22:01 crc kubenswrapper[4876]: I1205 09:22:01.338141 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-hxc5v_b2eec441-3198-440a-b061-47fe9bf9a340/kube-rbac-proxy/0.log" Dec 05 09:22:01 crc kubenswrapper[4876]: I1205 09:22:01.339927 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-hxc5v_b2eec441-3198-440a-b061-47fe9bf9a340/manager/0.log" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.562485 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8z82w"] Dec 05 09:22:17 crc kubenswrapper[4876]: E1205 09:22:17.563397 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" containerName="extract-utilities" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.563411 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" containerName="extract-utilities" Dec 05 09:22:17 crc kubenswrapper[4876]: E1205 09:22:17.563424 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="registry-server" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.563430 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="registry-server" Dec 05 09:22:17 crc kubenswrapper[4876]: E1205 09:22:17.563455 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" containerName="registry-server" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.563461 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" containerName="registry-server" Dec 05 09:22:17 crc kubenswrapper[4876]: E1205 09:22:17.563474 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="extract-utilities" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.563480 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="extract-utilities" Dec 05 09:22:17 crc kubenswrapper[4876]: E1205 09:22:17.563497 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" containerName="extract-content" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.563503 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" containerName="extract-content" Dec 05 09:22:17 crc kubenswrapper[4876]: E1205 09:22:17.563527 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="extract-content" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.563532 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="extract-content" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.563710 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="00ee21fe-67b7-42fc-84ee-cbe6ce5b500c" containerName="registry-server" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.563728 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="b09de4f9-07c4-41d6-801c-872380203cc5" containerName="registry-server" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.565343 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.586368 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z82w"] Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.627824 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-utilities\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.627872 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5b54\" (UniqueName: \"kubernetes.io/projected/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-kube-api-access-v5b54\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.627898 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-catalog-content\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.729487 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-utilities\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.729535 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5b54\" (UniqueName: \"kubernetes.io/projected/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-kube-api-access-v5b54\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.729566 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-catalog-content\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.730210 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-utilities\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.730229 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-catalog-content\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.752018 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5b54\" (UniqueName: \"kubernetes.io/projected/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-kube-api-access-v5b54\") pod \"redhat-marketplace-8z82w\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:17 crc kubenswrapper[4876]: I1205 09:22:17.934566 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:18 crc kubenswrapper[4876]: I1205 09:22:18.414256 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z82w"] Dec 05 09:22:19 crc kubenswrapper[4876]: I1205 09:22:19.299673 4876 generic.go:334] "Generic (PLEG): container finished" podID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerID="258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba" exitCode=0 Dec 05 09:22:19 crc kubenswrapper[4876]: I1205 09:22:19.300007 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z82w" event={"ID":"f89fa964-2b55-4bab-ba07-55ad6b77dcf0","Type":"ContainerDied","Data":"258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba"} Dec 05 09:22:19 crc kubenswrapper[4876]: I1205 09:22:19.300036 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z82w" event={"ID":"f89fa964-2b55-4bab-ba07-55ad6b77dcf0","Type":"ContainerStarted","Data":"229d6f8bc60c4c9ca643fd8cd38326f4692d034e90ee0c2294c44679438b63c7"} Dec 05 09:22:19 crc kubenswrapper[4876]: I1205 09:22:19.752801 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rbfjn_06abb292-ed23-4de1-9974-c487e7bb2933/control-plane-machine-set-operator/0.log" Dec 05 09:22:19 crc kubenswrapper[4876]: I1205 09:22:19.946142 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-w85pf_dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf/kube-rbac-proxy/0.log" Dec 05 09:22:20 crc kubenswrapper[4876]: I1205 09:22:20.005741 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-w85pf_dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf/machine-api-operator/0.log" Dec 05 09:22:20 crc kubenswrapper[4876]: I1205 09:22:20.312420 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z82w" event={"ID":"f89fa964-2b55-4bab-ba07-55ad6b77dcf0","Type":"ContainerStarted","Data":"329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425"} Dec 05 09:22:21 crc kubenswrapper[4876]: I1205 09:22:21.324944 4876 generic.go:334] "Generic (PLEG): container finished" podID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerID="329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425" exitCode=0 Dec 05 09:22:21 crc kubenswrapper[4876]: I1205 09:22:21.325024 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z82w" event={"ID":"f89fa964-2b55-4bab-ba07-55ad6b77dcf0","Type":"ContainerDied","Data":"329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425"} Dec 05 09:22:23 crc kubenswrapper[4876]: I1205 09:22:23.344353 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z82w" event={"ID":"f89fa964-2b55-4bab-ba07-55ad6b77dcf0","Type":"ContainerStarted","Data":"80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0"} Dec 05 09:22:23 crc kubenswrapper[4876]: I1205 09:22:23.378105 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8z82w" podStartSLOduration=2.9628094149999997 podStartE2EDuration="6.378088106s" podCreationTimestamp="2025-12-05 09:22:17 +0000 UTC" firstStartedPulling="2025-12-05 09:22:19.301718844 +0000 UTC m=+4003.790383456" lastFinishedPulling="2025-12-05 09:22:22.716997525 +0000 UTC m=+4007.205662147" observedRunningTime="2025-12-05 09:22:23.360289597 +0000 UTC m=+4007.848954219" watchObservedRunningTime="2025-12-05 09:22:23.378088106 +0000 UTC m=+4007.866752728" Dec 05 09:22:27 crc kubenswrapper[4876]: I1205 09:22:27.935315 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:27 crc kubenswrapper[4876]: I1205 09:22:27.935686 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:27 crc kubenswrapper[4876]: I1205 09:22:27.981700 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:28 crc kubenswrapper[4876]: I1205 09:22:28.434342 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:28 crc kubenswrapper[4876]: I1205 09:22:28.484554 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z82w"] Dec 05 09:22:30 crc kubenswrapper[4876]: I1205 09:22:30.407982 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8z82w" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerName="registry-server" containerID="cri-o://80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0" gracePeriod=2 Dec 05 09:22:30 crc kubenswrapper[4876]: I1205 09:22:30.940973 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.039495 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-catalog-content\") pod \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.040120 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-utilities\") pod \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.040335 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5b54\" (UniqueName: \"kubernetes.io/projected/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-kube-api-access-v5b54\") pod \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\" (UID: \"f89fa964-2b55-4bab-ba07-55ad6b77dcf0\") " Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.042810 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-utilities" (OuterVolumeSpecName: "utilities") pod "f89fa964-2b55-4bab-ba07-55ad6b77dcf0" (UID: "f89fa964-2b55-4bab-ba07-55ad6b77dcf0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.048082 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-kube-api-access-v5b54" (OuterVolumeSpecName: "kube-api-access-v5b54") pod "f89fa964-2b55-4bab-ba07-55ad6b77dcf0" (UID: "f89fa964-2b55-4bab-ba07-55ad6b77dcf0"). InnerVolumeSpecName "kube-api-access-v5b54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.063526 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f89fa964-2b55-4bab-ba07-55ad6b77dcf0" (UID: "f89fa964-2b55-4bab-ba07-55ad6b77dcf0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.142574 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.142621 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5b54\" (UniqueName: \"kubernetes.io/projected/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-kube-api-access-v5b54\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.142635 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89fa964-2b55-4bab-ba07-55ad6b77dcf0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.418982 4876 generic.go:334] "Generic (PLEG): container finished" podID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerID="80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0" exitCode=0 Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.419021 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z82w" event={"ID":"f89fa964-2b55-4bab-ba07-55ad6b77dcf0","Type":"ContainerDied","Data":"80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0"} Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.419054 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z82w" event={"ID":"f89fa964-2b55-4bab-ba07-55ad6b77dcf0","Type":"ContainerDied","Data":"229d6f8bc60c4c9ca643fd8cd38326f4692d034e90ee0c2294c44679438b63c7"} Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.419052 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8z82w" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.419087 4876 scope.go:117] "RemoveContainer" containerID="80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.448062 4876 scope.go:117] "RemoveContainer" containerID="329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.458562 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z82w"] Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.472815 4876 scope.go:117] "RemoveContainer" containerID="258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.473421 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z82w"] Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.531872 4876 scope.go:117] "RemoveContainer" containerID="80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0" Dec 05 09:22:31 crc kubenswrapper[4876]: E1205 09:22:31.532524 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0\": container with ID starting with 80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0 not found: ID does not exist" containerID="80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.532631 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0"} err="failed to get container status \"80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0\": rpc error: code = NotFound desc = could not find container \"80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0\": container with ID starting with 80f01523f0d67387bbc3c8058563e2ab02379cba8f2bf751f0d4b4937c525ea0 not found: ID does not exist" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.532772 4876 scope.go:117] "RemoveContainer" containerID="329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425" Dec 05 09:22:31 crc kubenswrapper[4876]: E1205 09:22:31.537387 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425\": container with ID starting with 329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425 not found: ID does not exist" containerID="329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.537532 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425"} err="failed to get container status \"329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425\": rpc error: code = NotFound desc = could not find container \"329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425\": container with ID starting with 329405f0c60b00ba4f3402f4c37759ab49e6eb7fcc308121353ac29f288c9425 not found: ID does not exist" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.537644 4876 scope.go:117] "RemoveContainer" containerID="258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba" Dec 05 09:22:31 crc kubenswrapper[4876]: E1205 09:22:31.541134 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba\": container with ID starting with 258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba not found: ID does not exist" containerID="258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.541168 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba"} err="failed to get container status \"258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba\": rpc error: code = NotFound desc = could not find container \"258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba\": container with ID starting with 258d257dc730355cb86c8c1182d7fd82c4ca00df61f0e77179ddb0f65bce05ba not found: ID does not exist" Dec 05 09:22:31 crc kubenswrapper[4876]: I1205 09:22:31.838798 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" path="/var/lib/kubelet/pods/f89fa964-2b55-4bab-ba07-55ad6b77dcf0/volumes" Dec 05 09:22:32 crc kubenswrapper[4876]: I1205 09:22:32.398855 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-qrhd7_ac3650c2-046a-4e3a-a5ce-f0b4b293efae/cert-manager-controller/0.log" Dec 05 09:22:32 crc kubenswrapper[4876]: I1205 09:22:32.500799 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-gwp9v_cdad8ccc-4995-4398-b83e-10a97b7187c9/cert-manager-cainjector/0.log" Dec 05 09:22:32 crc kubenswrapper[4876]: I1205 09:22:32.600222 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-clclm_398bf935-c5be-48a2-a68f-f884fa8dac57/cert-manager-webhook/0.log" Dec 05 09:22:44 crc kubenswrapper[4876]: I1205 09:22:44.211343 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-g4pnh_85c45836-2564-48e9-8c2f-33db89131c1b/nmstate-console-plugin/0.log" Dec 05 09:22:44 crc kubenswrapper[4876]: I1205 09:22:44.361349 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-n2snz_8c0975c5-f6cd-4565-a5fc-c9247bc8871c/nmstate-handler/0.log" Dec 05 09:22:44 crc kubenswrapper[4876]: I1205 09:22:44.380258 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-n4rqm_2fb0c525-e24c-4093-b8f7-8a9e3f788c29/kube-rbac-proxy/0.log" Dec 05 09:22:44 crc kubenswrapper[4876]: I1205 09:22:44.437498 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-n4rqm_2fb0c525-e24c-4093-b8f7-8a9e3f788c29/nmstate-metrics/0.log" Dec 05 09:22:44 crc kubenswrapper[4876]: I1205 09:22:44.555466 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-p68nr_9ef2d625-98dd-4299-bee0-848f9571b8c7/nmstate-operator/0.log" Dec 05 09:22:44 crc kubenswrapper[4876]: I1205 09:22:44.615906 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-dhgxp_f256549a-a4ac-40d4-b0b9-74e5b888d9f4/nmstate-webhook/0.log" Dec 05 09:22:59 crc kubenswrapper[4876]: I1205 09:22:59.902514 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-77dp9_460def3f-57ee-475e-bf28-788a977ea803/kube-rbac-proxy/0.log" Dec 05 09:22:59 crc kubenswrapper[4876]: I1205 09:22:59.987298 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-77dp9_460def3f-57ee-475e-bf28-788a977ea803/controller/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.156810 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-frr-files/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.345438 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-metrics/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.393036 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-frr-files/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.401188 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-reloader/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.440019 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-reloader/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.607045 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-reloader/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.624630 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-frr-files/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.654559 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-metrics/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.656128 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-metrics/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.928415 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-frr-files/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.970747 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-metrics/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.981563 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/controller/0.log" Dec 05 09:23:00 crc kubenswrapper[4876]: I1205 09:23:00.983563 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-reloader/0.log" Dec 05 09:23:01 crc kubenswrapper[4876]: I1205 09:23:01.212179 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/frr-metrics/0.log" Dec 05 09:23:01 crc kubenswrapper[4876]: I1205 09:23:01.268574 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/kube-rbac-proxy/0.log" Dec 05 09:23:01 crc kubenswrapper[4876]: I1205 09:23:01.298376 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/kube-rbac-proxy-frr/0.log" Dec 05 09:23:01 crc kubenswrapper[4876]: I1205 09:23:01.453276 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/reloader/0.log" Dec 05 09:23:01 crc kubenswrapper[4876]: I1205 09:23:01.527239 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-gvmp4_28f9db86-d329-4fa3-b808-b85aabf3840d/frr-k8s-webhook-server/0.log" Dec 05 09:23:01 crc kubenswrapper[4876]: I1205 09:23:01.718418 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-649974fd96-8cnpp_aea74793-6a5c-4aab-af14-8029af652439/manager/0.log" Dec 05 09:23:02 crc kubenswrapper[4876]: I1205 09:23:02.044727 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9k4hh_9a572583-7629-4436-b956-4b5befcb75ca/kube-rbac-proxy/0.log" Dec 05 09:23:02 crc kubenswrapper[4876]: I1205 09:23:02.091283 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b5765d7c-vzz6j_0710617c-73f2-44c9-a6fb-15f19c3a4aed/webhook-server/0.log" Dec 05 09:23:02 crc kubenswrapper[4876]: I1205 09:23:02.261184 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/frr/0.log" Dec 05 09:23:02 crc kubenswrapper[4876]: I1205 09:23:02.626948 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9k4hh_9a572583-7629-4436-b956-4b5befcb75ca/speaker/0.log" Dec 05 09:23:08 crc kubenswrapper[4876]: I1205 09:23:08.213935 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:23:08 crc kubenswrapper[4876]: I1205 09:23:08.215248 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:23:15 crc kubenswrapper[4876]: I1205 09:23:15.850827 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/util/0.log" Dec 05 09:23:15 crc kubenswrapper[4876]: I1205 09:23:15.987191 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/util/0.log" Dec 05 09:23:16 crc kubenswrapper[4876]: I1205 09:23:16.000521 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/pull/0.log" Dec 05 09:23:16 crc kubenswrapper[4876]: I1205 09:23:16.632381 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/pull/0.log" Dec 05 09:23:16 crc kubenswrapper[4876]: I1205 09:23:16.790884 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/util/0.log" Dec 05 09:23:16 crc kubenswrapper[4876]: I1205 09:23:16.805491 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/extract/0.log" Dec 05 09:23:16 crc kubenswrapper[4876]: I1205 09:23:16.805872 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/pull/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.009001 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/util/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.206945 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/util/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.215557 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/pull/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.280880 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/pull/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.395002 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/util/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.453282 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/pull/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.462938 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/extract/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.565067 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/util/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.767098 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/pull/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.777004 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/util/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.798099 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/pull/0.log" Dec 05 09:23:17 crc kubenswrapper[4876]: I1205 09:23:17.960716 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/pull/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.005523 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/util/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.017212 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/extract/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.179980 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-utilities/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.339204 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-utilities/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.366555 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-content/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.387017 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-content/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.706046 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-utilities/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.729457 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-content/0.log" Dec 05 09:23:18 crc kubenswrapper[4876]: I1205 09:23:18.961390 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-utilities/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.081504 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/registry-server/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.147424 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-utilities/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.205320 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-content/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.217312 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-content/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.428070 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-utilities/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.431875 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-content/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.486360 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5hd7v_e0282181-9f04-4c52-882f-322cd5a3ab11/marketplace-operator/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.639503 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-utilities/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.719025 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/registry-server/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.874753 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-utilities/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.879561 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-content/0.log" Dec 05 09:23:19 crc kubenswrapper[4876]: I1205 09:23:19.879560 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-content/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.071974 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-utilities/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.093944 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-content/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.111720 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-utilities/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.216914 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/registry-server/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.366563 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-content/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.366698 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-content/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.383240 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-utilities/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.504684 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-content/0.log" Dec 05 09:23:20 crc kubenswrapper[4876]: I1205 09:23:20.514500 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-utilities/0.log" Dec 05 09:23:21 crc kubenswrapper[4876]: I1205 09:23:21.123298 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/registry-server/0.log" Dec 05 09:23:35 crc kubenswrapper[4876]: I1205 09:23:35.883410 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-k9rt9_8c2118f9-ee5c-4fb7-a734-f9fbc098c53b/prometheus-operator/0.log" Dec 05 09:23:36 crc kubenswrapper[4876]: I1205 09:23:36.519157 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-644c864545-4vxch_84bb503d-0098-4342-bad3-6dc8e2ae325b/prometheus-operator-admission-webhook/0.log" Dec 05 09:23:36 crc kubenswrapper[4876]: I1205 09:23:36.519185 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-644c864545-c54zb_cc522201-89c0-478e-8e63-4962e162b4e6/prometheus-operator-admission-webhook/0.log" Dec 05 09:23:36 crc kubenswrapper[4876]: I1205 09:23:36.708818 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-v7c8l_714c939d-43ee-4564-851d-198cc78fbf7f/operator/0.log" Dec 05 09:23:36 crc kubenswrapper[4876]: I1205 09:23:36.772254 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-clc56_c5c1bf2d-26d2-4df3-8c0a-fe99d9715603/perses-operator/0.log" Dec 05 09:23:38 crc kubenswrapper[4876]: I1205 09:23:38.214192 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:23:38 crc kubenswrapper[4876]: I1205 09:23:38.214450 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:24:08 crc kubenswrapper[4876]: I1205 09:24:08.213763 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:24:08 crc kubenswrapper[4876]: I1205 09:24:08.214389 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:24:08 crc kubenswrapper[4876]: I1205 09:24:08.214433 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 09:24:08 crc kubenswrapper[4876]: I1205 09:24:08.215275 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6e7b471536bd5685c4904d6dc57a671c171f7553fd3d753fe9eabbc7e548d4f"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:24:08 crc kubenswrapper[4876]: I1205 09:24:08.215366 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://e6e7b471536bd5685c4904d6dc57a671c171f7553fd3d753fe9eabbc7e548d4f" gracePeriod=600 Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.412776 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="e6e7b471536bd5685c4904d6dc57a671c171f7553fd3d753fe9eabbc7e548d4f" exitCode=0 Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.413115 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"e6e7b471536bd5685c4904d6dc57a671c171f7553fd3d753fe9eabbc7e548d4f"} Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.413152 4876 scope.go:117] "RemoveContainer" containerID="ab2194c70386d314c9aa2ad32332c1740fa832d45177284a29b3ea589ed43101" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.775502 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rmc7p"] Dec 05 09:24:09 crc kubenswrapper[4876]: E1205 09:24:09.776415 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerName="registry-server" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.776437 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerName="registry-server" Dec 05 09:24:09 crc kubenswrapper[4876]: E1205 09:24:09.776495 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerName="extract-utilities" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.776504 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerName="extract-utilities" Dec 05 09:24:09 crc kubenswrapper[4876]: E1205 09:24:09.776526 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerName="extract-content" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.776535 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerName="extract-content" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.776826 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89fa964-2b55-4bab-ba07-55ad6b77dcf0" containerName="registry-server" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.779344 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.793355 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gwjf\" (UniqueName: \"kubernetes.io/projected/3958c99c-a229-47e0-9299-995c629de51a-kube-api-access-6gwjf\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.793433 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-catalog-content\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.793495 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-utilities\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.794911 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rmc7p"] Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.895214 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-utilities\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.895342 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gwjf\" (UniqueName: \"kubernetes.io/projected/3958c99c-a229-47e0-9299-995c629de51a-kube-api-access-6gwjf\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.895433 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-catalog-content\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.895999 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-catalog-content\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.896217 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-utilities\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:09 crc kubenswrapper[4876]: I1205 09:24:09.925286 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gwjf\" (UniqueName: \"kubernetes.io/projected/3958c99c-a229-47e0-9299-995c629de51a-kube-api-access-6gwjf\") pod \"certified-operators-rmc7p\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:10 crc kubenswrapper[4876]: I1205 09:24:10.109226 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:10 crc kubenswrapper[4876]: I1205 09:24:10.424497 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87"} Dec 05 09:24:11 crc kubenswrapper[4876]: I1205 09:24:11.100442 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rmc7p"] Dec 05 09:24:11 crc kubenswrapper[4876]: W1205 09:24:11.106591 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3958c99c_a229_47e0_9299_995c629de51a.slice/crio-c7abd262d50ceaaa69624cfe9ed83fbfaa91f2317fe4c9696962863de1d38a67 WatchSource:0}: Error finding container c7abd262d50ceaaa69624cfe9ed83fbfaa91f2317fe4c9696962863de1d38a67: Status 404 returned error can't find the container with id c7abd262d50ceaaa69624cfe9ed83fbfaa91f2317fe4c9696962863de1d38a67 Dec 05 09:24:11 crc kubenswrapper[4876]: I1205 09:24:11.435365 4876 generic.go:334] "Generic (PLEG): container finished" podID="3958c99c-a229-47e0-9299-995c629de51a" containerID="45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5" exitCode=0 Dec 05 09:24:11 crc kubenswrapper[4876]: I1205 09:24:11.435571 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmc7p" event={"ID":"3958c99c-a229-47e0-9299-995c629de51a","Type":"ContainerDied","Data":"45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5"} Dec 05 09:24:11 crc kubenswrapper[4876]: I1205 09:24:11.435688 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmc7p" event={"ID":"3958c99c-a229-47e0-9299-995c629de51a","Type":"ContainerStarted","Data":"c7abd262d50ceaaa69624cfe9ed83fbfaa91f2317fe4c9696962863de1d38a67"} Dec 05 09:24:14 crc kubenswrapper[4876]: I1205 09:24:14.476311 4876 generic.go:334] "Generic (PLEG): container finished" podID="3958c99c-a229-47e0-9299-995c629de51a" containerID="83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0" exitCode=0 Dec 05 09:24:14 crc kubenswrapper[4876]: I1205 09:24:14.476380 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmc7p" event={"ID":"3958c99c-a229-47e0-9299-995c629de51a","Type":"ContainerDied","Data":"83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0"} Dec 05 09:24:16 crc kubenswrapper[4876]: I1205 09:24:16.498853 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmc7p" event={"ID":"3958c99c-a229-47e0-9299-995c629de51a","Type":"ContainerStarted","Data":"e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1"} Dec 05 09:24:16 crc kubenswrapper[4876]: I1205 09:24:16.524412 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rmc7p" podStartSLOduration=3.060675233 podStartE2EDuration="7.524391124s" podCreationTimestamp="2025-12-05 09:24:09 +0000 UTC" firstStartedPulling="2025-12-05 09:24:11.437125323 +0000 UTC m=+4115.925789945" lastFinishedPulling="2025-12-05 09:24:15.900841224 +0000 UTC m=+4120.389505836" observedRunningTime="2025-12-05 09:24:16.513782529 +0000 UTC m=+4121.002447161" watchObservedRunningTime="2025-12-05 09:24:16.524391124 +0000 UTC m=+4121.013055746" Dec 05 09:24:20 crc kubenswrapper[4876]: I1205 09:24:20.110590 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:20 crc kubenswrapper[4876]: I1205 09:24:20.112270 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:20 crc kubenswrapper[4876]: I1205 09:24:20.175235 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:21 crc kubenswrapper[4876]: I1205 09:24:21.601332 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:21 crc kubenswrapper[4876]: I1205 09:24:21.654985 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rmc7p"] Dec 05 09:24:23 crc kubenswrapper[4876]: I1205 09:24:23.574454 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rmc7p" podUID="3958c99c-a229-47e0-9299-995c629de51a" containerName="registry-server" containerID="cri-o://e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1" gracePeriod=2 Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.016277 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.094536 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-utilities\") pod \"3958c99c-a229-47e0-9299-995c629de51a\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.095024 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gwjf\" (UniqueName: \"kubernetes.io/projected/3958c99c-a229-47e0-9299-995c629de51a-kube-api-access-6gwjf\") pod \"3958c99c-a229-47e0-9299-995c629de51a\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.095146 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-catalog-content\") pod \"3958c99c-a229-47e0-9299-995c629de51a\" (UID: \"3958c99c-a229-47e0-9299-995c629de51a\") " Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.096635 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-utilities" (OuterVolumeSpecName: "utilities") pod "3958c99c-a229-47e0-9299-995c629de51a" (UID: "3958c99c-a229-47e0-9299-995c629de51a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.102371 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3958c99c-a229-47e0-9299-995c629de51a-kube-api-access-6gwjf" (OuterVolumeSpecName: "kube-api-access-6gwjf") pod "3958c99c-a229-47e0-9299-995c629de51a" (UID: "3958c99c-a229-47e0-9299-995c629de51a"). InnerVolumeSpecName "kube-api-access-6gwjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.154824 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3958c99c-a229-47e0-9299-995c629de51a" (UID: "3958c99c-a229-47e0-9299-995c629de51a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.197388 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gwjf\" (UniqueName: \"kubernetes.io/projected/3958c99c-a229-47e0-9299-995c629de51a-kube-api-access-6gwjf\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.197435 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.197448 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3958c99c-a229-47e0-9299-995c629de51a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.587980 4876 generic.go:334] "Generic (PLEG): container finished" podID="3958c99c-a229-47e0-9299-995c629de51a" containerID="e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1" exitCode=0 Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.588045 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmc7p" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.588038 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmc7p" event={"ID":"3958c99c-a229-47e0-9299-995c629de51a","Type":"ContainerDied","Data":"e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1"} Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.588328 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmc7p" event={"ID":"3958c99c-a229-47e0-9299-995c629de51a","Type":"ContainerDied","Data":"c7abd262d50ceaaa69624cfe9ed83fbfaa91f2317fe4c9696962863de1d38a67"} Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.588350 4876 scope.go:117] "RemoveContainer" containerID="e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.622922 4876 scope.go:117] "RemoveContainer" containerID="83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.632651 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rmc7p"] Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.641919 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rmc7p"] Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.651077 4876 scope.go:117] "RemoveContainer" containerID="45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.697349 4876 scope.go:117] "RemoveContainer" containerID="e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1" Dec 05 09:24:24 crc kubenswrapper[4876]: E1205 09:24:24.699928 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1\": container with ID starting with e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1 not found: ID does not exist" containerID="e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.700003 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1"} err="failed to get container status \"e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1\": rpc error: code = NotFound desc = could not find container \"e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1\": container with ID starting with e73fc5ac8a536076c70450a23956a054df1196d83a7eb4ea1a8f6d97f74871a1 not found: ID does not exist" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.700034 4876 scope.go:117] "RemoveContainer" containerID="83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0" Dec 05 09:24:24 crc kubenswrapper[4876]: E1205 09:24:24.700355 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0\": container with ID starting with 83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0 not found: ID does not exist" containerID="83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.700387 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0"} err="failed to get container status \"83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0\": rpc error: code = NotFound desc = could not find container \"83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0\": container with ID starting with 83c88b471dccc4528e79166c8ccf214336e1276f27b099a83fee4cc4ffb016f0 not found: ID does not exist" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.700406 4876 scope.go:117] "RemoveContainer" containerID="45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5" Dec 05 09:24:24 crc kubenswrapper[4876]: E1205 09:24:24.700758 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5\": container with ID starting with 45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5 not found: ID does not exist" containerID="45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5" Dec 05 09:24:24 crc kubenswrapper[4876]: I1205 09:24:24.700783 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5"} err="failed to get container status \"45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5\": rpc error: code = NotFound desc = could not find container \"45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5\": container with ID starting with 45ca9a1bc648a7566602211dcd9df2e4212d2b5375a37faee3b46801957c0dc5 not found: ID does not exist" Dec 05 09:24:25 crc kubenswrapper[4876]: I1205 09:24:25.837112 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3958c99c-a229-47e0-9299-995c629de51a" path="/var/lib/kubelet/pods/3958c99c-a229-47e0-9299-995c629de51a/volumes" Dec 05 09:25:10 crc kubenswrapper[4876]: I1205 09:25:10.085341 4876 generic.go:334] "Generic (PLEG): container finished" podID="fcdc43e7-3f46-493c-963c-fa640144404e" containerID="452f47735e63f94a1ae2ccbc7d2c35e7f693033b26d771fc6035d34e0e9a9619" exitCode=0 Dec 05 09:25:10 crc kubenswrapper[4876]: I1205 09:25:10.085398 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kv4pt/must-gather-b24tw" event={"ID":"fcdc43e7-3f46-493c-963c-fa640144404e","Type":"ContainerDied","Data":"452f47735e63f94a1ae2ccbc7d2c35e7f693033b26d771fc6035d34e0e9a9619"} Dec 05 09:25:10 crc kubenswrapper[4876]: I1205 09:25:10.088224 4876 scope.go:117] "RemoveContainer" containerID="452f47735e63f94a1ae2ccbc7d2c35e7f693033b26d771fc6035d34e0e9a9619" Dec 05 09:25:10 crc kubenswrapper[4876]: I1205 09:25:10.455986 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kv4pt_must-gather-b24tw_fcdc43e7-3f46-493c-963c-fa640144404e/gather/0.log" Dec 05 09:25:18 crc kubenswrapper[4876]: I1205 09:25:18.972496 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kv4pt/must-gather-b24tw"] Dec 05 09:25:18 crc kubenswrapper[4876]: I1205 09:25:18.973351 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-kv4pt/must-gather-b24tw" podUID="fcdc43e7-3f46-493c-963c-fa640144404e" containerName="copy" containerID="cri-o://0e96b49683f81aa4b91ac7e72cdfedf421ebe4faf51aa5135aabecb7858bd3b7" gracePeriod=2 Dec 05 09:25:18 crc kubenswrapper[4876]: I1205 09:25:18.984774 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kv4pt/must-gather-b24tw"] Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.180435 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kv4pt_must-gather-b24tw_fcdc43e7-3f46-493c-963c-fa640144404e/copy/0.log" Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.181305 4876 generic.go:334] "Generic (PLEG): container finished" podID="fcdc43e7-3f46-493c-963c-fa640144404e" containerID="0e96b49683f81aa4b91ac7e72cdfedf421ebe4faf51aa5135aabecb7858bd3b7" exitCode=143 Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.416264 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kv4pt_must-gather-b24tw_fcdc43e7-3f46-493c-963c-fa640144404e/copy/0.log" Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.417419 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.443564 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fcdc43e7-3f46-493c-963c-fa640144404e-must-gather-output\") pod \"fcdc43e7-3f46-493c-963c-fa640144404e\" (UID: \"fcdc43e7-3f46-493c-963c-fa640144404e\") " Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.444017 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq2nh\" (UniqueName: \"kubernetes.io/projected/fcdc43e7-3f46-493c-963c-fa640144404e-kube-api-access-mq2nh\") pod \"fcdc43e7-3f46-493c-963c-fa640144404e\" (UID: \"fcdc43e7-3f46-493c-963c-fa640144404e\") " Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.459210 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcdc43e7-3f46-493c-963c-fa640144404e-kube-api-access-mq2nh" (OuterVolumeSpecName: "kube-api-access-mq2nh") pod "fcdc43e7-3f46-493c-963c-fa640144404e" (UID: "fcdc43e7-3f46-493c-963c-fa640144404e"). InnerVolumeSpecName "kube-api-access-mq2nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.546346 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq2nh\" (UniqueName: \"kubernetes.io/projected/fcdc43e7-3f46-493c-963c-fa640144404e-kube-api-access-mq2nh\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.603385 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcdc43e7-3f46-493c-963c-fa640144404e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fcdc43e7-3f46-493c-963c-fa640144404e" (UID: "fcdc43e7-3f46-493c-963c-fa640144404e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.649064 4876 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fcdc43e7-3f46-493c-963c-fa640144404e-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 05 09:25:19 crc kubenswrapper[4876]: I1205 09:25:19.836056 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcdc43e7-3f46-493c-963c-fa640144404e" path="/var/lib/kubelet/pods/fcdc43e7-3f46-493c-963c-fa640144404e/volumes" Dec 05 09:25:20 crc kubenswrapper[4876]: I1205 09:25:20.192730 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kv4pt_must-gather-b24tw_fcdc43e7-3f46-493c-963c-fa640144404e/copy/0.log" Dec 05 09:25:20 crc kubenswrapper[4876]: I1205 09:25:20.194219 4876 scope.go:117] "RemoveContainer" containerID="0e96b49683f81aa4b91ac7e72cdfedf421ebe4faf51aa5135aabecb7858bd3b7" Dec 05 09:25:20 crc kubenswrapper[4876]: I1205 09:25:20.194256 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kv4pt/must-gather-b24tw" Dec 05 09:25:20 crc kubenswrapper[4876]: I1205 09:25:20.214038 4876 scope.go:117] "RemoveContainer" containerID="452f47735e63f94a1ae2ccbc7d2c35e7f693033b26d771fc6035d34e0e9a9619" Dec 05 09:26:38 crc kubenswrapper[4876]: I1205 09:26:38.213341 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:26:38 crc kubenswrapper[4876]: I1205 09:26:38.213936 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:26:43 crc kubenswrapper[4876]: I1205 09:26:43.341447 4876 scope.go:117] "RemoveContainer" containerID="06c7e13827f955085ea031930c2ac21bd4e664534be26d6db495bbfe2059630b" Dec 05 09:27:08 crc kubenswrapper[4876]: I1205 09:27:08.213369 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:27:08 crc kubenswrapper[4876]: I1205 09:27:08.213959 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.213080 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.213537 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.213818 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.214591 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.214658 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" gracePeriod=600 Dec 05 09:27:38 crc kubenswrapper[4876]: E1205 09:27:38.576606 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.886506 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" exitCode=0 Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.886553 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87"} Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.886588 4876 scope.go:117] "RemoveContainer" containerID="e6e7b471536bd5685c4904d6dc57a671c171f7553fd3d753fe9eabbc7e548d4f" Dec 05 09:27:38 crc kubenswrapper[4876]: I1205 09:27:38.887332 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:27:38 crc kubenswrapper[4876]: E1205 09:27:38.887579 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:27:43 crc kubenswrapper[4876]: I1205 09:27:43.722581 4876 scope.go:117] "RemoveContainer" containerID="967082afac87fa43dec0b3297d2c3746207fa771135526cd2b29ed1511be6beb" Dec 05 09:27:43 crc kubenswrapper[4876]: I1205 09:27:43.750658 4876 scope.go:117] "RemoveContainer" containerID="e113f5cf129e4314f07f57a05f0cf3509efd5a331aff56037b1073fc0bbd323c" Dec 05 09:27:43 crc kubenswrapper[4876]: I1205 09:27:43.793394 4876 scope.go:117] "RemoveContainer" containerID="e3bdb6619e6883af1ec27ba33378979c71db2abe8c8828581396119183923322" Dec 05 09:27:53 crc kubenswrapper[4876]: I1205 09:27:53.824557 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:27:53 crc kubenswrapper[4876]: E1205 09:27:53.825502 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:28:05 crc kubenswrapper[4876]: I1205 09:28:05.834917 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:28:05 crc kubenswrapper[4876]: E1205 09:28:05.835825 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.294457 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-z2wdt/must-gather-cc4mq"] Dec 05 09:28:16 crc kubenswrapper[4876]: E1205 09:28:16.295359 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcdc43e7-3f46-493c-963c-fa640144404e" containerName="gather" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.295372 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcdc43e7-3f46-493c-963c-fa640144404e" containerName="gather" Dec 05 09:28:16 crc kubenswrapper[4876]: E1205 09:28:16.295386 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcdc43e7-3f46-493c-963c-fa640144404e" containerName="copy" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.295392 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcdc43e7-3f46-493c-963c-fa640144404e" containerName="copy" Dec 05 09:28:16 crc kubenswrapper[4876]: E1205 09:28:16.295423 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3958c99c-a229-47e0-9299-995c629de51a" containerName="extract-content" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.295429 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3958c99c-a229-47e0-9299-995c629de51a" containerName="extract-content" Dec 05 09:28:16 crc kubenswrapper[4876]: E1205 09:28:16.295439 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3958c99c-a229-47e0-9299-995c629de51a" containerName="extract-utilities" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.295444 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3958c99c-a229-47e0-9299-995c629de51a" containerName="extract-utilities" Dec 05 09:28:16 crc kubenswrapper[4876]: E1205 09:28:16.295455 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3958c99c-a229-47e0-9299-995c629de51a" containerName="registry-server" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.295461 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="3958c99c-a229-47e0-9299-995c629de51a" containerName="registry-server" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.295642 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="3958c99c-a229-47e0-9299-995c629de51a" containerName="registry-server" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.295657 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcdc43e7-3f46-493c-963c-fa640144404e" containerName="copy" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.295666 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcdc43e7-3f46-493c-963c-fa640144404e" containerName="gather" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.296842 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.310655 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-z2wdt"/"kube-root-ca.crt" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.310689 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-z2wdt"/"openshift-service-ca.crt" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.311126 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-z2wdt"/"default-dockercfg-jfx4b" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.313099 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-z2wdt/must-gather-cc4mq"] Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.376016 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljz45\" (UniqueName: \"kubernetes.io/projected/4a5a3968-3ce7-4022-8687-c9de4995e1fd-kube-api-access-ljz45\") pod \"must-gather-cc4mq\" (UID: \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\") " pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.376261 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4a5a3968-3ce7-4022-8687-c9de4995e1fd-must-gather-output\") pod \"must-gather-cc4mq\" (UID: \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\") " pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.478244 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4a5a3968-3ce7-4022-8687-c9de4995e1fd-must-gather-output\") pod \"must-gather-cc4mq\" (UID: \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\") " pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.478393 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljz45\" (UniqueName: \"kubernetes.io/projected/4a5a3968-3ce7-4022-8687-c9de4995e1fd-kube-api-access-ljz45\") pod \"must-gather-cc4mq\" (UID: \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\") " pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.478801 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4a5a3968-3ce7-4022-8687-c9de4995e1fd-must-gather-output\") pod \"must-gather-cc4mq\" (UID: \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\") " pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.500471 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljz45\" (UniqueName: \"kubernetes.io/projected/4a5a3968-3ce7-4022-8687-c9de4995e1fd-kube-api-access-ljz45\") pod \"must-gather-cc4mq\" (UID: \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\") " pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:28:16 crc kubenswrapper[4876]: I1205 09:28:16.625367 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:28:17 crc kubenswrapper[4876]: I1205 09:28:17.147435 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-z2wdt/must-gather-cc4mq"] Dec 05 09:28:17 crc kubenswrapper[4876]: I1205 09:28:17.249039 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" event={"ID":"4a5a3968-3ce7-4022-8687-c9de4995e1fd","Type":"ContainerStarted","Data":"8c4627a90816102d49cfd1c93f34416f0ce93fd54880d353c42fbfcde5532596"} Dec 05 09:28:17 crc kubenswrapper[4876]: I1205 09:28:17.823876 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:28:17 crc kubenswrapper[4876]: E1205 09:28:17.825459 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:28:18 crc kubenswrapper[4876]: I1205 09:28:18.271528 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" event={"ID":"4a5a3968-3ce7-4022-8687-c9de4995e1fd","Type":"ContainerStarted","Data":"882906be0595d14f56881d44a236e6b823622955bf329fa4c7c912439df0b67e"} Dec 05 09:28:18 crc kubenswrapper[4876]: I1205 09:28:18.271790 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" event={"ID":"4a5a3968-3ce7-4022-8687-c9de4995e1fd","Type":"ContainerStarted","Data":"e6ad7bce468b0e870c0afdae7ef3e84fa59faf2189f4983772ebeb68f825908e"} Dec 05 09:28:18 crc kubenswrapper[4876]: I1205 09:28:18.298407 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" podStartSLOduration=2.29839241 podStartE2EDuration="2.29839241s" podCreationTimestamp="2025-12-05 09:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:28:18.286514018 +0000 UTC m=+4362.775178640" watchObservedRunningTime="2025-12-05 09:28:18.29839241 +0000 UTC m=+4362.787057032" Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.233053 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-z2wdt/crc-debug-9hmv2"] Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.236221 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.320274 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85trk\" (UniqueName: \"kubernetes.io/projected/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-kube-api-access-85trk\") pod \"crc-debug-9hmv2\" (UID: \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\") " pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.320494 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-host\") pod \"crc-debug-9hmv2\" (UID: \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\") " pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.422387 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85trk\" (UniqueName: \"kubernetes.io/projected/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-kube-api-access-85trk\") pod \"crc-debug-9hmv2\" (UID: \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\") " pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.422927 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-host\") pod \"crc-debug-9hmv2\" (UID: \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\") " pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.423053 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-host\") pod \"crc-debug-9hmv2\" (UID: \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\") " pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.449116 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85trk\" (UniqueName: \"kubernetes.io/projected/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-kube-api-access-85trk\") pod \"crc-debug-9hmv2\" (UID: \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\") " pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:22 crc kubenswrapper[4876]: I1205 09:28:22.559921 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:22 crc kubenswrapper[4876]: W1205 09:28:22.590606 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4902c9e_e22b_4a0f_95d7_62072a4d94f1.slice/crio-78e136142d0d225d870ee9c35f82fa02d5714df59581f580c7f2f952a7f6f234 WatchSource:0}: Error finding container 78e136142d0d225d870ee9c35f82fa02d5714df59581f580c7f2f952a7f6f234: Status 404 returned error can't find the container with id 78e136142d0d225d870ee9c35f82fa02d5714df59581f580c7f2f952a7f6f234 Dec 05 09:28:23 crc kubenswrapper[4876]: I1205 09:28:23.592399 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" event={"ID":"e4902c9e-e22b-4a0f-95d7-62072a4d94f1","Type":"ContainerStarted","Data":"2d4276f6753b12ac4a50cced02f42741adb5f507a3ba72f712fefd9b7d85a21e"} Dec 05 09:28:23 crc kubenswrapper[4876]: I1205 09:28:23.592845 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" event={"ID":"e4902c9e-e22b-4a0f-95d7-62072a4d94f1","Type":"ContainerStarted","Data":"78e136142d0d225d870ee9c35f82fa02d5714df59581f580c7f2f952a7f6f234"} Dec 05 09:28:23 crc kubenswrapper[4876]: I1205 09:28:23.609822 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" podStartSLOduration=1.609805181 podStartE2EDuration="1.609805181s" podCreationTimestamp="2025-12-05 09:28:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:28:23.605808183 +0000 UTC m=+4368.094472815" watchObservedRunningTime="2025-12-05 09:28:23.609805181 +0000 UTC m=+4368.098469803" Dec 05 09:28:30 crc kubenswrapper[4876]: I1205 09:28:30.824403 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:28:30 crc kubenswrapper[4876]: E1205 09:28:30.825074 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:28:34 crc kubenswrapper[4876]: I1205 09:28:34.691669 4876 generic.go:334] "Generic (PLEG): container finished" podID="e4902c9e-e22b-4a0f-95d7-62072a4d94f1" containerID="2d4276f6753b12ac4a50cced02f42741adb5f507a3ba72f712fefd9b7d85a21e" exitCode=0 Dec 05 09:28:34 crc kubenswrapper[4876]: I1205 09:28:34.691718 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" event={"ID":"e4902c9e-e22b-4a0f-95d7-62072a4d94f1","Type":"ContainerDied","Data":"2d4276f6753b12ac4a50cced02f42741adb5f507a3ba72f712fefd9b7d85a21e"} Dec 05 09:28:35 crc kubenswrapper[4876]: I1205 09:28:35.838419 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:35 crc kubenswrapper[4876]: I1205 09:28:35.885526 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-z2wdt/crc-debug-9hmv2"] Dec 05 09:28:35 crc kubenswrapper[4876]: I1205 09:28:35.896570 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-z2wdt/crc-debug-9hmv2"] Dec 05 09:28:36 crc kubenswrapper[4876]: I1205 09:28:36.005883 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-host\") pod \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\" (UID: \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\") " Dec 05 09:28:36 crc kubenswrapper[4876]: I1205 09:28:36.005983 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-host" (OuterVolumeSpecName: "host") pod "e4902c9e-e22b-4a0f-95d7-62072a4d94f1" (UID: "e4902c9e-e22b-4a0f-95d7-62072a4d94f1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:28:36 crc kubenswrapper[4876]: I1205 09:28:36.006279 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85trk\" (UniqueName: \"kubernetes.io/projected/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-kube-api-access-85trk\") pod \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\" (UID: \"e4902c9e-e22b-4a0f-95d7-62072a4d94f1\") " Dec 05 09:28:36 crc kubenswrapper[4876]: I1205 09:28:36.008029 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-host\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:36 crc kubenswrapper[4876]: I1205 09:28:36.465886 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-kube-api-access-85trk" (OuterVolumeSpecName: "kube-api-access-85trk") pod "e4902c9e-e22b-4a0f-95d7-62072a4d94f1" (UID: "e4902c9e-e22b-4a0f-95d7-62072a4d94f1"). InnerVolumeSpecName "kube-api-access-85trk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:36 crc kubenswrapper[4876]: I1205 09:28:36.519101 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85trk\" (UniqueName: \"kubernetes.io/projected/e4902c9e-e22b-4a0f-95d7-62072a4d94f1-kube-api-access-85trk\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:36 crc kubenswrapper[4876]: I1205 09:28:36.713647 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78e136142d0d225d870ee9c35f82fa02d5714df59581f580c7f2f952a7f6f234" Dec 05 09:28:36 crc kubenswrapper[4876]: I1205 09:28:36.713710 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/crc-debug-9hmv2" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.688411 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-z2wdt/crc-debug-hp8z9"] Dec 05 09:28:37 crc kubenswrapper[4876]: E1205 09:28:37.688825 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4902c9e-e22b-4a0f-95d7-62072a4d94f1" containerName="container-00" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.688837 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4902c9e-e22b-4a0f-95d7-62072a4d94f1" containerName="container-00" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.689063 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4902c9e-e22b-4a0f-95d7-62072a4d94f1" containerName="container-00" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.689695 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.837434 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4902c9e-e22b-4a0f-95d7-62072a4d94f1" path="/var/lib/kubelet/pods/e4902c9e-e22b-4a0f-95d7-62072a4d94f1/volumes" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.844052 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjrph\" (UniqueName: \"kubernetes.io/projected/2a894c64-73f7-4500-971a-9a3b5e8a7f04-kube-api-access-pjrph\") pod \"crc-debug-hp8z9\" (UID: \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\") " pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.844172 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a894c64-73f7-4500-971a-9a3b5e8a7f04-host\") pod \"crc-debug-hp8z9\" (UID: \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\") " pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.946133 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a894c64-73f7-4500-971a-9a3b5e8a7f04-host\") pod \"crc-debug-hp8z9\" (UID: \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\") " pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.946353 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a894c64-73f7-4500-971a-9a3b5e8a7f04-host\") pod \"crc-debug-hp8z9\" (UID: \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\") " pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:37 crc kubenswrapper[4876]: I1205 09:28:37.946633 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjrph\" (UniqueName: \"kubernetes.io/projected/2a894c64-73f7-4500-971a-9a3b5e8a7f04-kube-api-access-pjrph\") pod \"crc-debug-hp8z9\" (UID: \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\") " pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:38 crc kubenswrapper[4876]: I1205 09:28:38.665723 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjrph\" (UniqueName: \"kubernetes.io/projected/2a894c64-73f7-4500-971a-9a3b5e8a7f04-kube-api-access-pjrph\") pod \"crc-debug-hp8z9\" (UID: \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\") " pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:38 crc kubenswrapper[4876]: I1205 09:28:38.921140 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:38 crc kubenswrapper[4876]: W1205 09:28:38.959785 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a894c64_73f7_4500_971a_9a3b5e8a7f04.slice/crio-c1e7e9a3a4b8c6e1b5dfc49fa06d903c4803a4f4adfc5e761214135ed07eacfb WatchSource:0}: Error finding container c1e7e9a3a4b8c6e1b5dfc49fa06d903c4803a4f4adfc5e761214135ed07eacfb: Status 404 returned error can't find the container with id c1e7e9a3a4b8c6e1b5dfc49fa06d903c4803a4f4adfc5e761214135ed07eacfb Dec 05 09:28:39 crc kubenswrapper[4876]: I1205 09:28:39.739974 4876 generic.go:334] "Generic (PLEG): container finished" podID="2a894c64-73f7-4500-971a-9a3b5e8a7f04" containerID="dfffe07a7e3ecb168e3b8565b97716b4b731cbf6532a89176427b61d182519af" exitCode=1 Dec 05 09:28:39 crc kubenswrapper[4876]: I1205 09:28:39.740284 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" event={"ID":"2a894c64-73f7-4500-971a-9a3b5e8a7f04","Type":"ContainerDied","Data":"dfffe07a7e3ecb168e3b8565b97716b4b731cbf6532a89176427b61d182519af"} Dec 05 09:28:39 crc kubenswrapper[4876]: I1205 09:28:39.740309 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" event={"ID":"2a894c64-73f7-4500-971a-9a3b5e8a7f04","Type":"ContainerStarted","Data":"c1e7e9a3a4b8c6e1b5dfc49fa06d903c4803a4f4adfc5e761214135ed07eacfb"} Dec 05 09:28:39 crc kubenswrapper[4876]: I1205 09:28:39.781878 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-z2wdt/crc-debug-hp8z9"] Dec 05 09:28:39 crc kubenswrapper[4876]: I1205 09:28:39.794919 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-z2wdt/crc-debug-hp8z9"] Dec 05 09:28:40 crc kubenswrapper[4876]: I1205 09:28:40.889139 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.021969 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a894c64-73f7-4500-971a-9a3b5e8a7f04-host\") pod \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\" (UID: \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\") " Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.022093 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a894c64-73f7-4500-971a-9a3b5e8a7f04-host" (OuterVolumeSpecName: "host") pod "2a894c64-73f7-4500-971a-9a3b5e8a7f04" (UID: "2a894c64-73f7-4500-971a-9a3b5e8a7f04"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.022224 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjrph\" (UniqueName: \"kubernetes.io/projected/2a894c64-73f7-4500-971a-9a3b5e8a7f04-kube-api-access-pjrph\") pod \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\" (UID: \"2a894c64-73f7-4500-971a-9a3b5e8a7f04\") " Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.022953 4876 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a894c64-73f7-4500-971a-9a3b5e8a7f04-host\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.031252 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a894c64-73f7-4500-971a-9a3b5e8a7f04-kube-api-access-pjrph" (OuterVolumeSpecName: "kube-api-access-pjrph") pod "2a894c64-73f7-4500-971a-9a3b5e8a7f04" (UID: "2a894c64-73f7-4500-971a-9a3b5e8a7f04"). InnerVolumeSpecName "kube-api-access-pjrph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.124145 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjrph\" (UniqueName: \"kubernetes.io/projected/2a894c64-73f7-4500-971a-9a3b5e8a7f04-kube-api-access-pjrph\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.759074 4876 scope.go:117] "RemoveContainer" containerID="dfffe07a7e3ecb168e3b8565b97716b4b731cbf6532a89176427b61d182519af" Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.759136 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/crc-debug-hp8z9" Dec 05 09:28:41 crc kubenswrapper[4876]: I1205 09:28:41.836323 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a894c64-73f7-4500-971a-9a3b5e8a7f04" path="/var/lib/kubelet/pods/2a894c64-73f7-4500-971a-9a3b5e8a7f04/volumes" Dec 05 09:28:42 crc kubenswrapper[4876]: I1205 09:28:42.823772 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:28:42 crc kubenswrapper[4876]: E1205 09:28:42.824135 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:28:55 crc kubenswrapper[4876]: I1205 09:28:55.833284 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:28:55 crc kubenswrapper[4876]: E1205 09:28:55.834142 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:29:07 crc kubenswrapper[4876]: I1205 09:29:07.824787 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:29:07 crc kubenswrapper[4876]: E1205 09:29:07.825483 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:29:20 crc kubenswrapper[4876]: I1205 09:29:20.824249 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:29:20 crc kubenswrapper[4876]: E1205 09:29:20.825132 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:29:32 crc kubenswrapper[4876]: I1205 09:29:32.824629 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:29:32 crc kubenswrapper[4876]: E1205 09:29:32.825517 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:29:45 crc kubenswrapper[4876]: I1205 09:29:45.844386 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:29:45 crc kubenswrapper[4876]: E1205 09:29:45.845232 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:29:46 crc kubenswrapper[4876]: I1205 09:29:46.794214 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a9faf08a-6b0e-435c-9810-d94212aa82ad/init-config-reloader/0.log" Dec 05 09:29:46 crc kubenswrapper[4876]: I1205 09:29:46.965387 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a9faf08a-6b0e-435c-9810-d94212aa82ad/init-config-reloader/0.log" Dec 05 09:29:46 crc kubenswrapper[4876]: I1205 09:29:46.967351 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a9faf08a-6b0e-435c-9810-d94212aa82ad/alertmanager/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.027938 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_a9faf08a-6b0e-435c-9810-d94212aa82ad/config-reloader/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.182307 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e960eaad-8be1-47d2-8e07-a17915f9cfea/aodh-evaluator/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.203221 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e960eaad-8be1-47d2-8e07-a17915f9cfea/aodh-listener/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.208197 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e960eaad-8be1-47d2-8e07-a17915f9cfea/aodh-api/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.311034 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_e960eaad-8be1-47d2-8e07-a17915f9cfea/aodh-notifier/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.430415 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9d7c76978-sl6zr_d9c4abfe-e5d6-4a04-88ff-3e721675a75e/barbican-api/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.441001 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9d7c76978-sl6zr_d9c4abfe-e5d6-4a04-88ff-3e721675a75e/barbican-api-log/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.580218 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-69575954c8-dwfz7_23f75a46-ea18-4217-bcb8-26d5652f95eb/barbican-keystone-listener/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.669409 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-69575954c8-dwfz7_23f75a46-ea18-4217-bcb8-26d5652f95eb/barbican-keystone-listener-log/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.786077 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f66c6dbdf-59mzz_aa8ee581-b9fc-4c2d-8139-b3d827f71697/barbican-worker/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.854260 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f66c6dbdf-59mzz_aa8ee581-b9fc-4c2d-8139-b3d827f71697/barbican-worker-log/0.log" Dec 05 09:29:47 crc kubenswrapper[4876]: I1205 09:29:47.978116 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-tg7jj_5f7a4eab-9eaf-4599-8eb6-deef38e971c9/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:48 crc kubenswrapper[4876]: I1205 09:29:48.621747 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2143f266-92fd-4c74-b305-0e1d95d3044b/ceilometer-central-agent/0.log" Dec 05 09:29:48 crc kubenswrapper[4876]: I1205 09:29:48.683068 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2143f266-92fd-4c74-b305-0e1d95d3044b/ceilometer-notification-agent/0.log" Dec 05 09:29:48 crc kubenswrapper[4876]: I1205 09:29:48.709723 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2143f266-92fd-4c74-b305-0e1d95d3044b/proxy-httpd/0.log" Dec 05 09:29:48 crc kubenswrapper[4876]: I1205 09:29:48.710007 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2143f266-92fd-4c74-b305-0e1d95d3044b/sg-core/0.log" Dec 05 09:29:48 crc kubenswrapper[4876]: I1205 09:29:48.892868 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9e321464-50bf-42c0-8584-a75e9282eb65/cinder-api/0.log" Dec 05 09:29:48 crc kubenswrapper[4876]: I1205 09:29:48.972927 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9e321464-50bf-42c0-8584-a75e9282eb65/cinder-api-log/0.log" Dec 05 09:29:49 crc kubenswrapper[4876]: I1205 09:29:49.114623 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ba584dcd-132e-407b-960e-4a68cfb1ea31/cinder-scheduler/0.log" Dec 05 09:29:49 crc kubenswrapper[4876]: I1205 09:29:49.183691 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_ba584dcd-132e-407b-960e-4a68cfb1ea31/probe/0.log" Dec 05 09:29:49 crc kubenswrapper[4876]: I1205 09:29:49.225068 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-pfqln_9d51e73b-805b-4690-aed5-514789933cce/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:49 crc kubenswrapper[4876]: I1205 09:29:49.429268 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-vpr4z_8a524784-c668-4388-81c5-719f97dbb947/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:49 crc kubenswrapper[4876]: I1205 09:29:49.456117 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-dcb46_ece6a347-25cd-41e9-89e0-f20c1fdb9eba/init/0.log" Dec 05 09:29:49 crc kubenswrapper[4876]: I1205 09:29:49.868941 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-dcb46_ece6a347-25cd-41e9-89e0-f20c1fdb9eba/init/0.log" Dec 05 09:29:49 crc kubenswrapper[4876]: I1205 09:29:49.912198 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-dcb46_ece6a347-25cd-41e9-89e0-f20c1fdb9eba/dnsmasq-dns/0.log" Dec 05 09:29:49 crc kubenswrapper[4876]: I1205 09:29:49.919603 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-gqzhn_d17d4141-7f2f-4598-9cf4-96870bca5903/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:50 crc kubenswrapper[4876]: I1205 09:29:50.092082 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e70a9dae-3695-42fa-bab0-b0ef8fc44e13/glance-httpd/0.log" Dec 05 09:29:50 crc kubenswrapper[4876]: I1205 09:29:50.124369 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_e70a9dae-3695-42fa-bab0-b0ef8fc44e13/glance-log/0.log" Dec 05 09:29:50 crc kubenswrapper[4876]: I1205 09:29:50.225629 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_00302ae2-a0ed-4a18-80e4-67ac2fed0230/glance-httpd/0.log" Dec 05 09:29:50 crc kubenswrapper[4876]: I1205 09:29:50.322028 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_00302ae2-a0ed-4a18-80e4-67ac2fed0230/glance-log/0.log" Dec 05 09:29:50 crc kubenswrapper[4876]: I1205 09:29:50.622388 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-b64b8b465-trj6k_6a133b19-0c3f-4a85-95ea-910b9c0c4df0/heat-engine/0.log" Dec 05 09:29:50 crc kubenswrapper[4876]: I1205 09:29:50.721160 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5f87bc649b-spcsf_b7673120-d39b-416e-8f5f-260cdce6f71b/heat-api/0.log" Dec 05 09:29:50 crc kubenswrapper[4876]: I1205 09:29:50.827169 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-lnm7b_ecc3ee00-cb0c-4084-ba46-80684e18f07e/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:50 crc kubenswrapper[4876]: I1205 09:29:50.882057 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-76699577f6-nt9wb_0f3a1341-ca1e-4b8c-a57d-0c83d39a403e/heat-cfnapi/0.log" Dec 05 09:29:51 crc kubenswrapper[4876]: I1205 09:29:51.061999 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-slf72_f596bae7-c567-469e-b7be-f95265c0f016/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:51 crc kubenswrapper[4876]: I1205 09:29:51.220974 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-577c476758-ksnrz_5612eb81-67eb-4559-b21a-549e4becca72/keystone-api/0.log" Dec 05 09:29:51 crc kubenswrapper[4876]: I1205 09:29:51.240413 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29415421-28x2l_778ede62-2e7e-4967-a96a-a8073b375f18/keystone-cron/0.log" Dec 05 09:29:51 crc kubenswrapper[4876]: I1205 09:29:51.410132 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1c06bc09-5dd5-4d2c-b572-078ea2a4d037/kube-state-metrics/0.log" Dec 05 09:29:51 crc kubenswrapper[4876]: I1205 09:29:51.448110 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-hwpsl_21e86dc4-3382-464d-aa9b-a811ee65e5d1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:51 crc kubenswrapper[4876]: I1205 09:29:51.778256 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5ff8f8c8c7-rx6w7_cd9bce2b-4ecb-45b0-a1b2-309c3f450a47/neutron-httpd/0.log" Dec 05 09:29:51 crc kubenswrapper[4876]: I1205 09:29:51.786336 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5ff8f8c8c7-rx6w7_cd9bce2b-4ecb-45b0-a1b2-309c3f450a47/neutron-api/0.log" Dec 05 09:29:51 crc kubenswrapper[4876]: I1205 09:29:51.987476 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-dq75c_74b443c1-b03a-42d2-8a66-49510fdfb5bb/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:52 crc kubenswrapper[4876]: I1205 09:29:52.188778 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b8b58af0-0308-45aa-825c-0d252a0bff0d/nova-api-log/0.log" Dec 05 09:29:52 crc kubenswrapper[4876]: I1205 09:29:52.657229 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_dc5cc0db-4ac5-4093-a0f0-abf202104b4f/nova-cell0-conductor-conductor/0.log" Dec 05 09:29:52 crc kubenswrapper[4876]: I1205 09:29:52.745091 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b8b58af0-0308-45aa-825c-0d252a0bff0d/nova-api-api/0.log" Dec 05 09:29:52 crc kubenswrapper[4876]: I1205 09:29:52.812482 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_b0643cc1-c10a-4da7-a3b8-09063ec41b80/nova-cell1-conductor-conductor/0.log" Dec 05 09:29:53 crc kubenswrapper[4876]: I1205 09:29:53.125987 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8ee1cf7e-9e7b-407a-ab48-d4abb9326b27/nova-cell1-novncproxy-novncproxy/0.log" Dec 05 09:29:53 crc kubenswrapper[4876]: I1205 09:29:53.150077 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5hxqz_d5afc65a-8adf-4aa6-a348-469c81b51bad/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:53 crc kubenswrapper[4876]: I1205 09:29:53.340399 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7ae3be31-2c00-4486-9c64-10c7fcb2c787/nova-metadata-log/0.log" Dec 05 09:29:53 crc kubenswrapper[4876]: I1205 09:29:53.624247 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_761374dd-700e-42fa-81d4-d560f97643be/mysql-bootstrap/0.log" Dec 05 09:29:53 crc kubenswrapper[4876]: I1205 09:29:53.664104 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_304bb281-a6d7-460a-a81e-3cecbb60e20f/nova-scheduler-scheduler/0.log" Dec 05 09:29:53 crc kubenswrapper[4876]: I1205 09:29:53.900227 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_761374dd-700e-42fa-81d4-d560f97643be/mysql-bootstrap/0.log" Dec 05 09:29:53 crc kubenswrapper[4876]: I1205 09:29:53.961490 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_761374dd-700e-42fa-81d4-d560f97643be/galera/0.log" Dec 05 09:29:54 crc kubenswrapper[4876]: I1205 09:29:54.111536 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d47eff06-2977-4bcb-bc3b-6ad08399bbd7/mysql-bootstrap/0.log" Dec 05 09:29:54 crc kubenswrapper[4876]: I1205 09:29:54.359685 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d47eff06-2977-4bcb-bc3b-6ad08399bbd7/mysql-bootstrap/0.log" Dec 05 09:29:54 crc kubenswrapper[4876]: I1205 09:29:54.431021 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d47eff06-2977-4bcb-bc3b-6ad08399bbd7/galera/0.log" Dec 05 09:29:54 crc kubenswrapper[4876]: I1205 09:29:54.617016 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_17ad89b3-21e5-4db2-8d47-29248aa9e054/openstackclient/0.log" Dec 05 09:29:54 crc kubenswrapper[4876]: I1205 09:29:54.729173 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-7p5wk_442faaa2-6ae4-4b4a-aad9-fa3f8f157ba2/ovn-controller/0.log" Dec 05 09:29:54 crc kubenswrapper[4876]: I1205 09:29:54.971747 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_7ae3be31-2c00-4486-9c64-10c7fcb2c787/nova-metadata-metadata/0.log" Dec 05 09:29:55 crc kubenswrapper[4876]: I1205 09:29:55.190460 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hxnnn_82469e38-4872-4d09-8a1e-2e635df62905/openstack-network-exporter/0.log" Dec 05 09:29:55 crc kubenswrapper[4876]: I1205 09:29:55.625105 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhwc7_07839a94-e94b-4ff2-9eaf-9fb315c83046/ovsdb-server-init/0.log" Dec 05 09:29:55 crc kubenswrapper[4876]: I1205 09:29:55.836397 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhwc7_07839a94-e94b-4ff2-9eaf-9fb315c83046/ovs-vswitchd/0.log" Dec 05 09:29:55 crc kubenswrapper[4876]: I1205 09:29:55.904541 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhwc7_07839a94-e94b-4ff2-9eaf-9fb315c83046/ovsdb-server-init/0.log" Dec 05 09:29:55 crc kubenswrapper[4876]: I1205 09:29:55.974603 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-fhwc7_07839a94-e94b-4ff2-9eaf-9fb315c83046/ovsdb-server/0.log" Dec 05 09:29:56 crc kubenswrapper[4876]: I1205 09:29:56.144495 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-pdx5k_33523fba-8378-49ee-b023-0f80a4a8f91b/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:56 crc kubenswrapper[4876]: I1205 09:29:56.245024 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f6758b3-0cba-4d9b-b3ad-25149d206ff9/openstack-network-exporter/0.log" Dec 05 09:29:56 crc kubenswrapper[4876]: I1205 09:29:56.275117 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f6758b3-0cba-4d9b-b3ad-25149d206ff9/ovn-northd/0.log" Dec 05 09:29:56 crc kubenswrapper[4876]: I1205 09:29:56.461029 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3f18961c-c654-4f94-98a3-697644fdfda4/openstack-network-exporter/0.log" Dec 05 09:29:56 crc kubenswrapper[4876]: I1205 09:29:56.497257 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3f18961c-c654-4f94-98a3-697644fdfda4/ovsdbserver-nb/0.log" Dec 05 09:29:56 crc kubenswrapper[4876]: I1205 09:29:56.631379 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9c3699c2-facd-4d82-92ec-4816b871833b/openstack-network-exporter/0.log" Dec 05 09:29:56 crc kubenswrapper[4876]: I1205 09:29:56.736319 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9c3699c2-facd-4d82-92ec-4816b871833b/ovsdbserver-sb/0.log" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.320888 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-56d57865bd-mfljr_27487f6b-feb8-47c1-98db-5ed28120b3f3/placement-api/0.log" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.331328 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-56d57865bd-mfljr_27487f6b-feb8-47c1-98db-5ed28120b3f3/placement-log/0.log" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.481848 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/init-config-reloader/0.log" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.700245 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/config-reloader/0.log" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.721593 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/init-config-reloader/0.log" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.725425 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/thanos-sidecar/0.log" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.827016 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_ff40adb3-f509-46ed-96f3-c915e75906b0/prometheus/0.log" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.831060 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:29:57 crc kubenswrapper[4876]: E1205 09:29:57.831325 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:29:57 crc kubenswrapper[4876]: I1205 09:29:57.986142 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ef895f6-3d13-4980-98ef-9111438fe1ec/setup-container/0.log" Dec 05 09:29:58 crc kubenswrapper[4876]: I1205 09:29:58.195518 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ef895f6-3d13-4980-98ef-9111438fe1ec/setup-container/0.log" Dec 05 09:29:58 crc kubenswrapper[4876]: I1205 09:29:58.207611 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_692469d9-b150-4d67-b412-ba8760598f8b/setup-container/0.log" Dec 05 09:29:58 crc kubenswrapper[4876]: I1205 09:29:58.227518 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_6ef895f6-3d13-4980-98ef-9111438fe1ec/rabbitmq/0.log" Dec 05 09:29:58 crc kubenswrapper[4876]: I1205 09:29:58.482784 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_692469d9-b150-4d67-b412-ba8760598f8b/setup-container/0.log" Dec 05 09:29:58 crc kubenswrapper[4876]: I1205 09:29:58.529759 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_692469d9-b150-4d67-b412-ba8760598f8b/rabbitmq/0.log" Dec 05 09:29:58 crc kubenswrapper[4876]: I1205 09:29:58.606754 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-gqx8w_a4528785-cb5d-4281-9de7-2140359d7c0c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:58 crc kubenswrapper[4876]: I1205 09:29:58.833224 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-7xhsh_fde6a633-2ee9-4fad-b157-24435a2e89d4/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:58 crc kubenswrapper[4876]: I1205 09:29:58.886541 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-hgjsz_4554e5a3-3a60-4bc6-85c7-bfdc19421f05/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:59 crc kubenswrapper[4876]: I1205 09:29:59.075321 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-8w987_87e85de2-b888-48c3-8f2d-e500ccedca3e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:29:59 crc kubenswrapper[4876]: I1205 09:29:59.218057 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-stmmf_dc2909c6-f489-467c-a203-01adb27428f5/ssh-known-hosts-edpm-deployment/0.log" Dec 05 09:29:59 crc kubenswrapper[4876]: I1205 09:29:59.460172 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-98b56b8f5-ksg4z_fbd7f2e5-b830-45be-be0b-1f23806634e2/proxy-server/0.log" Dec 05 09:29:59 crc kubenswrapper[4876]: I1205 09:29:59.543219 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-98b56b8f5-ksg4z_fbd7f2e5-b830-45be-be0b-1f23806634e2/proxy-httpd/0.log" Dec 05 09:29:59 crc kubenswrapper[4876]: I1205 09:29:59.562431 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-hzjsf_d41c7aad-b564-4c13-9f60-2d1fe53861a8/swift-ring-rebalance/0.log" Dec 05 09:29:59 crc kubenswrapper[4876]: I1205 09:29:59.783562 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/account-auditor/0.log" Dec 05 09:29:59 crc kubenswrapper[4876]: I1205 09:29:59.827616 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/account-reaper/0.log" Dec 05 09:29:59 crc kubenswrapper[4876]: I1205 09:29:59.906588 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/account-replicator/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.005716 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/container-auditor/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.035801 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/account-server/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.064647 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/container-replicator/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.219254 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw"] Dec 05 09:30:00 crc kubenswrapper[4876]: E1205 09:30:00.219799 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a894c64-73f7-4500-971a-9a3b5e8a7f04" containerName="container-00" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.219821 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a894c64-73f7-4500-971a-9a3b5e8a7f04" containerName="container-00" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.220117 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a894c64-73f7-4500-971a-9a3b5e8a7f04" containerName="container-00" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.223562 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.228321 4876 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.228627 4876 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.245062 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw"] Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.281077 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/container-updater/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.286422 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/container-server/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.297124 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m46bf\" (UniqueName: \"kubernetes.io/projected/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-kube-api-access-m46bf\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.297178 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-config-volume\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.297242 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-secret-volume\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.305925 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-auditor/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.357314 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-expirer/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.399354 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m46bf\" (UniqueName: \"kubernetes.io/projected/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-kube-api-access-m46bf\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.399397 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-config-volume\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.399431 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-secret-volume\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.400690 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-config-volume\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.406931 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-secret-volume\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.426247 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m46bf\" (UniqueName: \"kubernetes.io/projected/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-kube-api-access-m46bf\") pod \"collect-profiles-29415450-87ppw\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.462673 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-server/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.500614 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-replicator/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.525858 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/object-updater/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.564547 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.611071 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/rsync/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.805257 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_71735818-3684-481f-bbb3-72aff2d50428/swift-recon-cron/0.log" Dec 05 09:30:00 crc kubenswrapper[4876]: I1205 09:30:00.939132 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-zc47n_611e9ca4-b50e-415e-bf76-ec70f386af72/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:30:01 crc kubenswrapper[4876]: I1205 09:30:01.134128 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw"] Dec 05 09:30:01 crc kubenswrapper[4876]: I1205 09:30:01.141300 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-77bdc_91a239cd-2756-4d03-8175-32dd507c0b2a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 09:30:01 crc kubenswrapper[4876]: I1205 09:30:01.925887 4876 generic.go:334] "Generic (PLEG): container finished" podID="bd1e0057-b872-44c7-b7ff-9a97928e4cd5" containerID="86a93b840ce1a44ad5a85168c7a4162502ab591289f33f3e1d20db70e8c9527a" exitCode=0 Dec 05 09:30:01 crc kubenswrapper[4876]: I1205 09:30:01.926408 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" event={"ID":"bd1e0057-b872-44c7-b7ff-9a97928e4cd5","Type":"ContainerDied","Data":"86a93b840ce1a44ad5a85168c7a4162502ab591289f33f3e1d20db70e8c9527a"} Dec 05 09:30:01 crc kubenswrapper[4876]: I1205 09:30:01.926479 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" event={"ID":"bd1e0057-b872-44c7-b7ff-9a97928e4cd5","Type":"ContainerStarted","Data":"0a5411fa4e6d86cc121fd505074f72c773b298e0abb4b6673c438559bddf894d"} Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.441812 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.540076 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-config-volume\") pod \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.540269 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m46bf\" (UniqueName: \"kubernetes.io/projected/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-kube-api-access-m46bf\") pod \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.540322 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-secret-volume\") pod \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\" (UID: \"bd1e0057-b872-44c7-b7ff-9a97928e4cd5\") " Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.560115 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-kube-api-access-m46bf" (OuterVolumeSpecName: "kube-api-access-m46bf") pod "bd1e0057-b872-44c7-b7ff-9a97928e4cd5" (UID: "bd1e0057-b872-44c7-b7ff-9a97928e4cd5"). InnerVolumeSpecName "kube-api-access-m46bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.560579 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bd1e0057-b872-44c7-b7ff-9a97928e4cd5" (UID: "bd1e0057-b872-44c7-b7ff-9a97928e4cd5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.560919 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-config-volume" (OuterVolumeSpecName: "config-volume") pod "bd1e0057-b872-44c7-b7ff-9a97928e4cd5" (UID: "bd1e0057-b872-44c7-b7ff-9a97928e4cd5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.711330 4876 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.711680 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m46bf\" (UniqueName: \"kubernetes.io/projected/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-kube-api-access-m46bf\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.711694 4876 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd1e0057-b872-44c7-b7ff-9a97928e4cd5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.972799 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" event={"ID":"bd1e0057-b872-44c7-b7ff-9a97928e4cd5","Type":"ContainerDied","Data":"0a5411fa4e6d86cc121fd505074f72c773b298e0abb4b6673c438559bddf894d"} Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.972835 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a5411fa4e6d86cc121fd505074f72c773b298e0abb4b6673c438559bddf894d" Dec 05 09:30:03 crc kubenswrapper[4876]: I1205 09:30:03.972885 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-87ppw" Dec 05 09:30:04 crc kubenswrapper[4876]: I1205 09:30:04.534992 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq"] Dec 05 09:30:04 crc kubenswrapper[4876]: I1205 09:30:04.546290 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415405-2hpzq"] Dec 05 09:30:05 crc kubenswrapper[4876]: I1205 09:30:05.842581 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c" path="/var/lib/kubelet/pods/e84cc2fa-f9fb-49fc-a1cb-185ed872fa7c/volumes" Dec 05 09:30:09 crc kubenswrapper[4876]: I1205 09:30:09.823918 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:30:09 crc kubenswrapper[4876]: E1205 09:30:09.824810 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:30:12 crc kubenswrapper[4876]: I1205 09:30:12.501086 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_846c521d-f570-4731-a923-8a1e9c626659/memcached/0.log" Dec 05 09:30:23 crc kubenswrapper[4876]: I1205 09:30:23.826317 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:30:23 crc kubenswrapper[4876]: E1205 09:30:23.827117 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:30:29 crc kubenswrapper[4876]: I1205 09:30:29.787171 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/util/0.log" Dec 05 09:30:29 crc kubenswrapper[4876]: I1205 09:30:29.979471 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/util/0.log" Dec 05 09:30:29 crc kubenswrapper[4876]: I1205 09:30:29.991791 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/pull/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.017204 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/pull/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.157027 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/pull/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.165203 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/util/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.165342 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_40aff819d769a828833aec0cf8a0af26b654d3c6362fb8108b8d17d2ad6p8c4_01744036-ed48-4c66-b33a-a7fb0d95ab80/extract/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.316187 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ppsdr_04e3edd8-f29a-48d1-84f4-81985747740f/kube-rbac-proxy/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.384622 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ppsdr_04e3edd8-f29a-48d1-84f4-81985747740f/manager/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.465006 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-d4l6m_0036dfa3-4545-43f5-a6f6-c329485b678b/kube-rbac-proxy/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.589695 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-d4l6m_0036dfa3-4545-43f5-a6f6-c329485b678b/manager/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.620100 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-jwn22_d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca/kube-rbac-proxy/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.697321 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-jwn22_d1b1b3f4-28b5-44b8-a499-81e3fd64c8ca/manager/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.823549 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-bv2mb_1c4ffc06-424a-4c94-be0d-2ff27bf8d920/kube-rbac-proxy/0.log" Dec 05 09:30:30 crc kubenswrapper[4876]: I1205 09:30:30.937109 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-bv2mb_1c4ffc06-424a-4c94-be0d-2ff27bf8d920/manager/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.077763 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-q7r4r_291c6269-dc8a-4193-a602-2a16202db817/kube-rbac-proxy/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.199466 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-q7r4r_291c6269-dc8a-4193-a602-2a16202db817/manager/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.236646 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-vjh2v_0b8ca916-a8c5-4866-8333-ebbc0f861161/kube-rbac-proxy/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.295403 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-vjh2v_0b8ca916-a8c5-4866-8333-ebbc0f861161/manager/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.439103 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-l9jr7_679c172a-10ff-4a0e-8932-bcc1f78393d7/kube-rbac-proxy/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.635813 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-l9jr7_679c172a-10ff-4a0e-8932-bcc1f78393d7/manager/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.670733 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-qgxmr_ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7/kube-rbac-proxy/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.671794 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-qgxmr_ccc6d8d6-723f-4fb4-b92b-bd562c0b3aa7/manager/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.839307 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4ckhg_e21b6aa8-f703-42e0-8251-06444e6423fe/kube-rbac-proxy/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.938145 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4ckhg_e21b6aa8-f703-42e0-8251-06444e6423fe/manager/0.log" Dec 05 09:30:31 crc kubenswrapper[4876]: I1205 09:30:31.973834 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-v2p7b_9f436ce1-a9d9-411a-824a-85afc3ee2069/kube-rbac-proxy/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.061520 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-v2p7b_9f436ce1-a9d9-411a-824a-85afc3ee2069/manager/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.120707 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vp2x4_13b56ee6-b8fa-40bf-add8-1778f4acf30d/kube-rbac-proxy/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.178014 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-vp2x4_13b56ee6-b8fa-40bf-add8-1778f4acf30d/manager/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.369541 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lxwt9_7b0b34b8-da13-4978-9a68-606135b29834/manager/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.370857 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lxwt9_7b0b34b8-da13-4978-9a68-606135b29834/kube-rbac-proxy/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.503551 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2spcs_68006b91-3398-4ff1-8c6e-ede02835c0ef/kube-rbac-proxy/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.666119 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2spcs_68006b91-3398-4ff1-8c6e-ede02835c0ef/manager/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.760077 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-svc6v_fb571bcf-872c-42f5-9a65-18c38291b062/manager/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.765483 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-svc6v_fb571bcf-872c-42f5-9a65-18c38291b062/kube-rbac-proxy/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.871056 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6_e3a38502-6fad-4383-81e8-591318a4d4d3/kube-rbac-proxy/0.log" Dec 05 09:30:32 crc kubenswrapper[4876]: I1205 09:30:32.964522 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4k28v6_e3a38502-6fad-4383-81e8-591318a4d4d3/manager/0.log" Dec 05 09:30:33 crc kubenswrapper[4876]: I1205 09:30:33.326548 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qjjkm_437c91ca-0472-481b-b58f-40269f4d2c18/registry-server/0.log" Dec 05 09:30:33 crc kubenswrapper[4876]: I1205 09:30:33.546567 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-cc6d987f6-whr8k_1c603b75-2c70-4ac9-ba31-7f2f69785482/operator/0.log" Dec 05 09:30:33 crc kubenswrapper[4876]: I1205 09:30:33.585646 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lfncq_217cf4c3-b49c-4bc6-a1cd-a12589264d11/kube-rbac-proxy/0.log" Dec 05 09:30:33 crc kubenswrapper[4876]: I1205 09:30:33.780757 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lfncq_217cf4c3-b49c-4bc6-a1cd-a12589264d11/manager/0.log" Dec 05 09:30:33 crc kubenswrapper[4876]: I1205 09:30:33.819802 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wptqj_ca537166-1c8e-4616-ba9f-b97382f11d7f/kube-rbac-proxy/0.log" Dec 05 09:30:33 crc kubenswrapper[4876]: I1205 09:30:33.898639 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wptqj_ca537166-1c8e-4616-ba9f-b97382f11d7f/manager/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.103637 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qltmw_861cead7-82be-495f-b144-7120ee08ae2d/kube-rbac-proxy/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.105802 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-9ptpl_af4995ab-bc40-4446-bc2e-d421fe03423c/operator/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.236878 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-qltmw_861cead7-82be-495f-b144-7120ee08ae2d/manager/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.340101 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/kube-rbac-proxy/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.428073 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8579f747db-9t4h4_48cca8bf-c0e1-4957-885a-4cf42c1c33fe/manager/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.552423 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-59c7b9d69d-t9vv2_74207150-aafb-409d-b5fb-7d56b9ac8f9c/manager/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.558963 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tn7mh_af99aad7-cbab-4990-8ed5-3a66c9f4b432/kube-rbac-proxy/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.592993 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-tn7mh_af99aad7-cbab-4990-8ed5-3a66c9f4b432/manager/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.680235 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-hxc5v_b2eec441-3198-440a-b061-47fe9bf9a340/kube-rbac-proxy/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.689706 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-hxc5v_b2eec441-3198-440a-b061-47fe9bf9a340/manager/0.log" Dec 05 09:30:34 crc kubenswrapper[4876]: I1205 09:30:34.824579 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:30:34 crc kubenswrapper[4876]: E1205 09:30:34.824846 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:30:43 crc kubenswrapper[4876]: I1205 09:30:43.891199 4876 scope.go:117] "RemoveContainer" containerID="d1782f0f06de456c949ae455429bbd98e9c267814cd57f9f7a5bdc239badfb1c" Dec 05 09:30:46 crc kubenswrapper[4876]: I1205 09:30:46.824297 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:30:46 crc kubenswrapper[4876]: E1205 09:30:46.825128 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:30:54 crc kubenswrapper[4876]: I1205 09:30:54.778998 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rbfjn_06abb292-ed23-4de1-9974-c487e7bb2933/control-plane-machine-set-operator/0.log" Dec 05 09:30:54 crc kubenswrapper[4876]: I1205 09:30:54.963491 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-w85pf_dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf/kube-rbac-proxy/0.log" Dec 05 09:30:54 crc kubenswrapper[4876]: I1205 09:30:54.970284 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-w85pf_dbb73726-8f4a-4d39-b22f-2eb5c3dd5bdf/machine-api-operator/0.log" Dec 05 09:30:59 crc kubenswrapper[4876]: I1205 09:30:59.824131 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:30:59 crc kubenswrapper[4876]: E1205 09:30:59.825105 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:31:09 crc kubenswrapper[4876]: I1205 09:31:09.413337 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-qrhd7_ac3650c2-046a-4e3a-a5ce-f0b4b293efae/cert-manager-controller/0.log" Dec 05 09:31:09 crc kubenswrapper[4876]: I1205 09:31:09.422611 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-gwp9v_cdad8ccc-4995-4398-b83e-10a97b7187c9/cert-manager-cainjector/0.log" Dec 05 09:31:09 crc kubenswrapper[4876]: I1205 09:31:09.576633 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-clclm_398bf935-c5be-48a2-a68f-f884fa8dac57/cert-manager-webhook/0.log" Dec 05 09:31:14 crc kubenswrapper[4876]: I1205 09:31:14.824924 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:31:14 crc kubenswrapper[4876]: E1205 09:31:14.825714 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.211581 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nj7bw"] Dec 05 09:31:20 crc kubenswrapper[4876]: E1205 09:31:20.212690 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd1e0057-b872-44c7-b7ff-9a97928e4cd5" containerName="collect-profiles" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.212707 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd1e0057-b872-44c7-b7ff-9a97928e4cd5" containerName="collect-profiles" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.212992 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd1e0057-b872-44c7-b7ff-9a97928e4cd5" containerName="collect-profiles" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.214883 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.225361 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nj7bw"] Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.254334 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffmks\" (UniqueName: \"kubernetes.io/projected/51fa92df-ab98-43a6-9b77-ae5a85f84e36-kube-api-access-ffmks\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.254628 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-catalog-content\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.254851 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-utilities\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.357167 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-utilities\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.357372 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffmks\" (UniqueName: \"kubernetes.io/projected/51fa92df-ab98-43a6-9b77-ae5a85f84e36-kube-api-access-ffmks\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.357417 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-catalog-content\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.357736 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-utilities\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.357935 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-catalog-content\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.399636 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffmks\" (UniqueName: \"kubernetes.io/projected/51fa92df-ab98-43a6-9b77-ae5a85f84e36-kube-api-access-ffmks\") pod \"redhat-operators-nj7bw\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:20 crc kubenswrapper[4876]: I1205 09:31:20.571665 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:21 crc kubenswrapper[4876]: I1205 09:31:21.068693 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nj7bw"] Dec 05 09:31:21 crc kubenswrapper[4876]: I1205 09:31:21.570221 4876 generic.go:334] "Generic (PLEG): container finished" podID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerID="ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5" exitCode=0 Dec 05 09:31:21 crc kubenswrapper[4876]: I1205 09:31:21.570311 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj7bw" event={"ID":"51fa92df-ab98-43a6-9b77-ae5a85f84e36","Type":"ContainerDied","Data":"ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5"} Dec 05 09:31:21 crc kubenswrapper[4876]: I1205 09:31:21.570491 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj7bw" event={"ID":"51fa92df-ab98-43a6-9b77-ae5a85f84e36","Type":"ContainerStarted","Data":"cbbeae2fd751cff47f7d8dbafa14a6359cc0d49c8fed13ece2544800991a439e"} Dec 05 09:31:21 crc kubenswrapper[4876]: I1205 09:31:21.572609 4876 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:31:22 crc kubenswrapper[4876]: I1205 09:31:22.582009 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj7bw" event={"ID":"51fa92df-ab98-43a6-9b77-ae5a85f84e36","Type":"ContainerStarted","Data":"5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c"} Dec 05 09:31:22 crc kubenswrapper[4876]: I1205 09:31:22.923080 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-g4pnh_85c45836-2564-48e9-8c2f-33db89131c1b/nmstate-console-plugin/0.log" Dec 05 09:31:23 crc kubenswrapper[4876]: I1205 09:31:23.178748 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-n2snz_8c0975c5-f6cd-4565-a5fc-c9247bc8871c/nmstate-handler/0.log" Dec 05 09:31:23 crc kubenswrapper[4876]: I1205 09:31:23.259100 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-n4rqm_2fb0c525-e24c-4093-b8f7-8a9e3f788c29/nmstate-metrics/0.log" Dec 05 09:31:23 crc kubenswrapper[4876]: I1205 09:31:23.275221 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-n4rqm_2fb0c525-e24c-4093-b8f7-8a9e3f788c29/kube-rbac-proxy/0.log" Dec 05 09:31:23 crc kubenswrapper[4876]: I1205 09:31:23.376192 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-p68nr_9ef2d625-98dd-4299-bee0-848f9571b8c7/nmstate-operator/0.log" Dec 05 09:31:23 crc kubenswrapper[4876]: I1205 09:31:23.477440 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-dhgxp_f256549a-a4ac-40d4-b0b9-74e5b888d9f4/nmstate-webhook/0.log" Dec 05 09:31:26 crc kubenswrapper[4876]: I1205 09:31:26.637884 4876 generic.go:334] "Generic (PLEG): container finished" podID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerID="5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c" exitCode=0 Dec 05 09:31:26 crc kubenswrapper[4876]: I1205 09:31:26.637932 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj7bw" event={"ID":"51fa92df-ab98-43a6-9b77-ae5a85f84e36","Type":"ContainerDied","Data":"5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c"} Dec 05 09:31:26 crc kubenswrapper[4876]: I1205 09:31:26.823558 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:31:26 crc kubenswrapper[4876]: E1205 09:31:26.824310 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:31:27 crc kubenswrapper[4876]: I1205 09:31:27.648656 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj7bw" event={"ID":"51fa92df-ab98-43a6-9b77-ae5a85f84e36","Type":"ContainerStarted","Data":"88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904"} Dec 05 09:31:27 crc kubenswrapper[4876]: I1205 09:31:27.674152 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nj7bw" podStartSLOduration=2.159791095 podStartE2EDuration="7.674128468s" podCreationTimestamp="2025-12-05 09:31:20 +0000 UTC" firstStartedPulling="2025-12-05 09:31:21.572311042 +0000 UTC m=+4546.060975664" lastFinishedPulling="2025-12-05 09:31:27.086648405 +0000 UTC m=+4551.575313037" observedRunningTime="2025-12-05 09:31:27.667647782 +0000 UTC m=+4552.156312424" watchObservedRunningTime="2025-12-05 09:31:27.674128468 +0000 UTC m=+4552.162793090" Dec 05 09:31:30 crc kubenswrapper[4876]: I1205 09:31:30.572280 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:30 crc kubenswrapper[4876]: I1205 09:31:30.572580 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:31 crc kubenswrapper[4876]: I1205 09:31:31.620072 4876 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nj7bw" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="registry-server" probeResult="failure" output=< Dec 05 09:31:31 crc kubenswrapper[4876]: timeout: failed to connect service ":50051" within 1s Dec 05 09:31:31 crc kubenswrapper[4876]: > Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.084337 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6rlrk"] Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.086456 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.099141 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rlrk"] Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.258464 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-utilities\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.258551 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpb7t\" (UniqueName: \"kubernetes.io/projected/de75f30c-84a7-49d8-9e16-1761c433d0b9-kube-api-access-hpb7t\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.258649 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-catalog-content\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.360193 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-catalog-content\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.360389 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpb7t\" (UniqueName: \"kubernetes.io/projected/de75f30c-84a7-49d8-9e16-1761c433d0b9-kube-api-access-hpb7t\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.360415 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-utilities\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.360830 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-utilities\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.361061 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-catalog-content\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.388146 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpb7t\" (UniqueName: \"kubernetes.io/projected/de75f30c-84a7-49d8-9e16-1761c433d0b9-kube-api-access-hpb7t\") pod \"community-operators-6rlrk\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:32 crc kubenswrapper[4876]: I1205 09:31:32.412362 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:33 crc kubenswrapper[4876]: I1205 09:31:33.074133 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rlrk"] Dec 05 09:31:33 crc kubenswrapper[4876]: I1205 09:31:33.715291 4876 generic.go:334] "Generic (PLEG): container finished" podID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerID="2ff864a2a824f5ba4815dc605a057eed7e7751ab10214f2e33023b132f2d9383" exitCode=0 Dec 05 09:31:33 crc kubenswrapper[4876]: I1205 09:31:33.715393 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rlrk" event={"ID":"de75f30c-84a7-49d8-9e16-1761c433d0b9","Type":"ContainerDied","Data":"2ff864a2a824f5ba4815dc605a057eed7e7751ab10214f2e33023b132f2d9383"} Dec 05 09:31:33 crc kubenswrapper[4876]: I1205 09:31:33.715570 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rlrk" event={"ID":"de75f30c-84a7-49d8-9e16-1761c433d0b9","Type":"ContainerStarted","Data":"b151fc605e7fa5da513bbd4d25624691891ce413df530009861027764696795b"} Dec 05 09:31:34 crc kubenswrapper[4876]: I1205 09:31:34.730446 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rlrk" event={"ID":"de75f30c-84a7-49d8-9e16-1761c433d0b9","Type":"ContainerStarted","Data":"669e2b585ea3ff53e1be542acd0ac4c63b80325dab094d86959dc944ac1bd70d"} Dec 05 09:31:35 crc kubenswrapper[4876]: I1205 09:31:35.740098 4876 generic.go:334] "Generic (PLEG): container finished" podID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerID="669e2b585ea3ff53e1be542acd0ac4c63b80325dab094d86959dc944ac1bd70d" exitCode=0 Dec 05 09:31:35 crc kubenswrapper[4876]: I1205 09:31:35.740426 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rlrk" event={"ID":"de75f30c-84a7-49d8-9e16-1761c433d0b9","Type":"ContainerDied","Data":"669e2b585ea3ff53e1be542acd0ac4c63b80325dab094d86959dc944ac1bd70d"} Dec 05 09:31:36 crc kubenswrapper[4876]: I1205 09:31:36.751322 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rlrk" event={"ID":"de75f30c-84a7-49d8-9e16-1761c433d0b9","Type":"ContainerStarted","Data":"743122f3451ef8f38e713938bbe9eeefeca015e182433af12855d2b2d0b02f4f"} Dec 05 09:31:36 crc kubenswrapper[4876]: I1205 09:31:36.779352 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6rlrk" podStartSLOduration=2.377998131 podStartE2EDuration="4.779333659s" podCreationTimestamp="2025-12-05 09:31:32 +0000 UTC" firstStartedPulling="2025-12-05 09:31:33.717010852 +0000 UTC m=+4558.205675474" lastFinishedPulling="2025-12-05 09:31:36.11834638 +0000 UTC m=+4560.607011002" observedRunningTime="2025-12-05 09:31:36.768832624 +0000 UTC m=+4561.257497246" watchObservedRunningTime="2025-12-05 09:31:36.779333659 +0000 UTC m=+4561.267998281" Dec 05 09:31:40 crc kubenswrapper[4876]: I1205 09:31:40.621020 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:40 crc kubenswrapper[4876]: I1205 09:31:40.675048 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:40 crc kubenswrapper[4876]: I1205 09:31:40.880455 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nj7bw"] Dec 05 09:31:41 crc kubenswrapper[4876]: I1205 09:31:41.518130 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-77dp9_460def3f-57ee-475e-bf28-788a977ea803/kube-rbac-proxy/0.log" Dec 05 09:31:41 crc kubenswrapper[4876]: I1205 09:31:41.731425 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-77dp9_460def3f-57ee-475e-bf28-788a977ea803/controller/0.log" Dec 05 09:31:41 crc kubenswrapper[4876]: I1205 09:31:41.799993 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nj7bw" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="registry-server" containerID="cri-o://88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904" gracePeriod=2 Dec 05 09:31:41 crc kubenswrapper[4876]: I1205 09:31:41.817513 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-frr-files/0.log" Dec 05 09:31:41 crc kubenswrapper[4876]: I1205 09:31:41.825733 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:31:41 crc kubenswrapper[4876]: E1205 09:31:41.826305 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:31:41 crc kubenswrapper[4876]: I1205 09:31:41.941591 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-reloader/0.log" Dec 05 09:31:41 crc kubenswrapper[4876]: I1205 09:31:41.952416 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-frr-files/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.047709 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-metrics/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.108800 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-reloader/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.251335 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-frr-files/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.309376 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.331582 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-reloader/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.355272 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-metrics/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.379949 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-metrics/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.413438 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.413486 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.485645 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-utilities\") pod \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.485794 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffmks\" (UniqueName: \"kubernetes.io/projected/51fa92df-ab98-43a6-9b77-ae5a85f84e36-kube-api-access-ffmks\") pod \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.485877 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-catalog-content\") pod \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\" (UID: \"51fa92df-ab98-43a6-9b77-ae5a85f84e36\") " Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.486920 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-utilities" (OuterVolumeSpecName: "utilities") pod "51fa92df-ab98-43a6-9b77-ae5a85f84e36" (UID: "51fa92df-ab98-43a6-9b77-ae5a85f84e36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.487722 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.508489 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51fa92df-ab98-43a6-9b77-ae5a85f84e36-kube-api-access-ffmks" (OuterVolumeSpecName: "kube-api-access-ffmks") pod "51fa92df-ab98-43a6-9b77-ae5a85f84e36" (UID: "51fa92df-ab98-43a6-9b77-ae5a85f84e36"). InnerVolumeSpecName "kube-api-access-ffmks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.588703 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.588741 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffmks\" (UniqueName: \"kubernetes.io/projected/51fa92df-ab98-43a6-9b77-ae5a85f84e36-kube-api-access-ffmks\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.594447 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51fa92df-ab98-43a6-9b77-ae5a85f84e36" (UID: "51fa92df-ab98-43a6-9b77-ae5a85f84e36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.611457 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-reloader/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.629913 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-frr-files/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.637656 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/controller/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.645573 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/cp-metrics/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.689655 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51fa92df-ab98-43a6-9b77-ae5a85f84e36-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.805068 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/kube-rbac-proxy/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.812204 4876 generic.go:334] "Generic (PLEG): container finished" podID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerID="88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904" exitCode=0 Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.812268 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nj7bw" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.812300 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj7bw" event={"ID":"51fa92df-ab98-43a6-9b77-ae5a85f84e36","Type":"ContainerDied","Data":"88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904"} Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.812363 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nj7bw" event={"ID":"51fa92df-ab98-43a6-9b77-ae5a85f84e36","Type":"ContainerDied","Data":"cbbeae2fd751cff47f7d8dbafa14a6359cc0d49c8fed13ece2544800991a439e"} Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.812388 4876 scope.go:117] "RemoveContainer" containerID="88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.845189 4876 scope.go:117] "RemoveContainer" containerID="5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.852550 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/frr-metrics/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.853330 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/kube-rbac-proxy-frr/0.log" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.854747 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nj7bw"] Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.866443 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nj7bw"] Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.880707 4876 scope.go:117] "RemoveContainer" containerID="ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.880760 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.937239 4876 scope.go:117] "RemoveContainer" containerID="88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904" Dec 05 09:31:42 crc kubenswrapper[4876]: E1205 09:31:42.939181 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904\": container with ID starting with 88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904 not found: ID does not exist" containerID="88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.939210 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904"} err="failed to get container status \"88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904\": rpc error: code = NotFound desc = could not find container \"88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904\": container with ID starting with 88d7d74783840b26bff700cb410a56d1cd247a0d38f2b9adf954a704e3ac2904 not found: ID does not exist" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.939227 4876 scope.go:117] "RemoveContainer" containerID="5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c" Dec 05 09:31:42 crc kubenswrapper[4876]: E1205 09:31:42.939455 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c\": container with ID starting with 5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c not found: ID does not exist" containerID="5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.939473 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c"} err="failed to get container status \"5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c\": rpc error: code = NotFound desc = could not find container \"5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c\": container with ID starting with 5e9794925ea4a9cf99b0295a74ca74c371f6be4f65dd9c848d0ae46bffdc075c not found: ID does not exist" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.939484 4876 scope.go:117] "RemoveContainer" containerID="ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5" Dec 05 09:31:42 crc kubenswrapper[4876]: E1205 09:31:42.940242 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5\": container with ID starting with ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5 not found: ID does not exist" containerID="ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5" Dec 05 09:31:42 crc kubenswrapper[4876]: I1205 09:31:42.940285 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5"} err="failed to get container status \"ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5\": rpc error: code = NotFound desc = could not find container \"ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5\": container with ID starting with ceafeab76b4620edfa618706423cc0c6faa5a93607782d8b092bcf50c7545bd5 not found: ID does not exist" Dec 05 09:31:43 crc kubenswrapper[4876]: I1205 09:31:43.068133 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/reloader/0.log" Dec 05 09:31:43 crc kubenswrapper[4876]: I1205 09:31:43.111309 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-gvmp4_28f9db86-d329-4fa3-b808-b85aabf3840d/frr-k8s-webhook-server/0.log" Dec 05 09:31:43 crc kubenswrapper[4876]: I1205 09:31:43.360783 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-649974fd96-8cnpp_aea74793-6a5c-4aab-af14-8029af652439/manager/0.log" Dec 05 09:31:43 crc kubenswrapper[4876]: I1205 09:31:43.633074 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b5765d7c-vzz6j_0710617c-73f2-44c9-a6fb-15f19c3a4aed/webhook-server/0.log" Dec 05 09:31:43 crc kubenswrapper[4876]: I1205 09:31:43.836692 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" path="/var/lib/kubelet/pods/51fa92df-ab98-43a6-9b77-ae5a85f84e36/volumes" Dec 05 09:31:43 crc kubenswrapper[4876]: I1205 09:31:43.874596 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9k4hh_9a572583-7629-4436-b956-4b5befcb75ca/kube-rbac-proxy/0.log" Dec 05 09:31:44 crc kubenswrapper[4876]: I1205 09:31:44.523712 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9k4hh_9a572583-7629-4436-b956-4b5befcb75ca/speaker/0.log" Dec 05 09:31:44 crc kubenswrapper[4876]: I1205 09:31:44.541122 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fgbfx_2cdf28f4-aee9-424f-81b7-36e131a81f30/frr/0.log" Dec 05 09:31:44 crc kubenswrapper[4876]: I1205 09:31:44.876469 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rlrk"] Dec 05 09:31:44 crc kubenswrapper[4876]: I1205 09:31:44.877077 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6rlrk" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerName="registry-server" containerID="cri-o://743122f3451ef8f38e713938bbe9eeefeca015e182433af12855d2b2d0b02f4f" gracePeriod=2 Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.843309 4876 generic.go:334] "Generic (PLEG): container finished" podID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerID="743122f3451ef8f38e713938bbe9eeefeca015e182433af12855d2b2d0b02f4f" exitCode=0 Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.843625 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rlrk" event={"ID":"de75f30c-84a7-49d8-9e16-1761c433d0b9","Type":"ContainerDied","Data":"743122f3451ef8f38e713938bbe9eeefeca015e182433af12855d2b2d0b02f4f"} Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.843654 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rlrk" event={"ID":"de75f30c-84a7-49d8-9e16-1761c433d0b9","Type":"ContainerDied","Data":"b151fc605e7fa5da513bbd4d25624691891ce413df530009861027764696795b"} Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.843669 4876 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b151fc605e7fa5da513bbd4d25624691891ce413df530009861027764696795b" Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.908381 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.959885 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-catalog-content\") pod \"de75f30c-84a7-49d8-9e16-1761c433d0b9\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.960257 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-utilities\") pod \"de75f30c-84a7-49d8-9e16-1761c433d0b9\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.960596 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpb7t\" (UniqueName: \"kubernetes.io/projected/de75f30c-84a7-49d8-9e16-1761c433d0b9-kube-api-access-hpb7t\") pod \"de75f30c-84a7-49d8-9e16-1761c433d0b9\" (UID: \"de75f30c-84a7-49d8-9e16-1761c433d0b9\") " Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.960991 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-utilities" (OuterVolumeSpecName: "utilities") pod "de75f30c-84a7-49d8-9e16-1761c433d0b9" (UID: "de75f30c-84a7-49d8-9e16-1761c433d0b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.961284 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:45 crc kubenswrapper[4876]: I1205 09:31:45.965654 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de75f30c-84a7-49d8-9e16-1761c433d0b9-kube-api-access-hpb7t" (OuterVolumeSpecName: "kube-api-access-hpb7t") pod "de75f30c-84a7-49d8-9e16-1761c433d0b9" (UID: "de75f30c-84a7-49d8-9e16-1761c433d0b9"). InnerVolumeSpecName "kube-api-access-hpb7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:46 crc kubenswrapper[4876]: I1205 09:31:46.014393 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de75f30c-84a7-49d8-9e16-1761c433d0b9" (UID: "de75f30c-84a7-49d8-9e16-1761c433d0b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:46 crc kubenswrapper[4876]: I1205 09:31:46.065086 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpb7t\" (UniqueName: \"kubernetes.io/projected/de75f30c-84a7-49d8-9e16-1761c433d0b9-kube-api-access-hpb7t\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:46 crc kubenswrapper[4876]: I1205 09:31:46.065120 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de75f30c-84a7-49d8-9e16-1761c433d0b9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:46 crc kubenswrapper[4876]: I1205 09:31:46.852705 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rlrk" Dec 05 09:31:46 crc kubenswrapper[4876]: I1205 09:31:46.895219 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rlrk"] Dec 05 09:31:46 crc kubenswrapper[4876]: I1205 09:31:46.932620 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6rlrk"] Dec 05 09:31:47 crc kubenswrapper[4876]: I1205 09:31:47.836611 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" path="/var/lib/kubelet/pods/de75f30c-84a7-49d8-9e16-1761c433d0b9/volumes" Dec 05 09:31:56 crc kubenswrapper[4876]: I1205 09:31:56.824119 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:31:56 crc kubenswrapper[4876]: E1205 09:31:56.824886 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:31:58 crc kubenswrapper[4876]: I1205 09:31:58.371959 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/util/0.log" Dec 05 09:31:58 crc kubenswrapper[4876]: I1205 09:31:58.579808 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/pull/0.log" Dec 05 09:31:58 crc kubenswrapper[4876]: I1205 09:31:58.591512 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/pull/0.log" Dec 05 09:31:58 crc kubenswrapper[4876]: I1205 09:31:58.596949 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/util/0.log" Dec 05 09:31:58 crc kubenswrapper[4876]: I1205 09:31:58.746579 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/util/0.log" Dec 05 09:31:58 crc kubenswrapper[4876]: I1205 09:31:58.746601 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/extract/0.log" Dec 05 09:31:58 crc kubenswrapper[4876]: I1205 09:31:58.753148 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fk9qrt_42f3432a-0028-45dc-a0d3-a886bd2da16a/pull/0.log" Dec 05 09:31:58 crc kubenswrapper[4876]: I1205 09:31:58.916094 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/util/0.log" Dec 05 09:31:59 crc kubenswrapper[4876]: I1205 09:31:59.065973 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/pull/0.log" Dec 05 09:31:59 crc kubenswrapper[4876]: I1205 09:31:59.083377 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/util/0.log" Dec 05 09:31:59 crc kubenswrapper[4876]: I1205 09:31:59.134466 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/pull/0.log" Dec 05 09:31:59 crc kubenswrapper[4876]: I1205 09:31:59.300668 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/util/0.log" Dec 05 09:31:59 crc kubenswrapper[4876]: I1205 09:31:59.312315 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/pull/0.log" Dec 05 09:31:59 crc kubenswrapper[4876]: I1205 09:31:59.330327 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210kwmgv_1ee6bebd-088e-47ce-97a1-78d3132184f4/extract/0.log" Dec 05 09:31:59 crc kubenswrapper[4876]: I1205 09:31:59.490613 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/util/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.129352 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/util/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.139085 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/pull/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.156208 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/pull/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.290782 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/pull/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.299618 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/extract/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.339553 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835wvmd_41d46d23-6d8e-44e3-ba8e-9be76b08bff6/util/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.489034 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-utilities/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.630546 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-content/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.647444 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-content/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.657456 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-utilities/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.887193 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-utilities/0.log" Dec 05 09:32:00 crc kubenswrapper[4876]: I1205 09:32:00.953096 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/extract-content/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.169415 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-utilities/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.348106 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jlxfs_f48b9ece-bd4d-4f71-8238-99c5fdc0727a/registry-server/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.385641 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-utilities/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.406144 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-content/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.420927 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-content/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.667481 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-utilities/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.731569 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/extract-content/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.892469 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5hd7v_e0282181-9f04-4c52-882f-322cd5a3ab11/marketplace-operator/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.951215 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4sj77_062e963d-bece-4a6c-aa68-90648656b0d0/registry-server/0.log" Dec 05 09:32:01 crc kubenswrapper[4876]: I1205 09:32:01.989723 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-utilities/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.130447 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-utilities/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.175652 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-content/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.198317 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-content/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.374205 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-content/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.377221 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/extract-utilities/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.522289 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-utilities/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.531554 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8n6mr_97e7ccac-0d35-45a2-a36c-974f008cf0cc/registry-server/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.617718 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-content/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.626784 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-utilities/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.710713 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-content/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.863631 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-content/0.log" Dec 05 09:32:02 crc kubenswrapper[4876]: I1205 09:32:02.870233 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/extract-utilities/0.log" Dec 05 09:32:03 crc kubenswrapper[4876]: I1205 09:32:03.468797 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-89hsw_3df610b8-24ff-45eb-92cd-8fd40cd53f0c/registry-server/0.log" Dec 05 09:32:08 crc kubenswrapper[4876]: I1205 09:32:08.824538 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:32:08 crc kubenswrapper[4876]: E1205 09:32:08.826429 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:32:15 crc kubenswrapper[4876]: I1205 09:32:15.075868 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-k9rt9_8c2118f9-ee5c-4fb7-a734-f9fbc098c53b/prometheus-operator/0.log" Dec 05 09:32:15 crc kubenswrapper[4876]: I1205 09:32:15.352484 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-644c864545-c54zb_cc522201-89c0-478e-8e63-4962e162b4e6/prometheus-operator-admission-webhook/0.log" Dec 05 09:32:15 crc kubenswrapper[4876]: I1205 09:32:15.353079 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-644c864545-4vxch_84bb503d-0098-4342-bad3-6dc8e2ae325b/prometheus-operator-admission-webhook/0.log" Dec 05 09:32:15 crc kubenswrapper[4876]: I1205 09:32:15.547323 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-v7c8l_714c939d-43ee-4564-851d-198cc78fbf7f/operator/0.log" Dec 05 09:32:15 crc kubenswrapper[4876]: I1205 09:32:15.603232 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-clc56_c5c1bf2d-26d2-4df3-8c0a-fe99d9715603/perses-operator/0.log" Dec 05 09:32:21 crc kubenswrapper[4876]: I1205 09:32:21.826470 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:32:21 crc kubenswrapper[4876]: E1205 09:32:21.827121 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:32:36 crc kubenswrapper[4876]: I1205 09:32:36.825717 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:32:36 crc kubenswrapper[4876]: E1205 09:32:36.826519 4876 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xj8qx_openshift-machine-config-operator(77322cc8-c6ab-4250-8098-9938309f0af8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" Dec 05 09:32:51 crc kubenswrapper[4876]: I1205 09:32:51.824036 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:32:52 crc kubenswrapper[4876]: I1205 09:32:52.484909 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"5c754475cbb8158c1c8983f993d9812c384247e50a836b7438f9ae73893ac0eb"} Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.940593 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-79kf4"] Dec 05 09:33:02 crc kubenswrapper[4876]: E1205 09:33:02.941620 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerName="extract-utilities" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.941637 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerName="extract-utilities" Dec 05 09:33:02 crc kubenswrapper[4876]: E1205 09:33:02.941660 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="registry-server" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.941669 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="registry-server" Dec 05 09:33:02 crc kubenswrapper[4876]: E1205 09:33:02.941682 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="extract-content" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.941691 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="extract-content" Dec 05 09:33:02 crc kubenswrapper[4876]: E1205 09:33:02.941705 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="extract-utilities" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.941712 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="extract-utilities" Dec 05 09:33:02 crc kubenswrapper[4876]: E1205 09:33:02.941723 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerName="extract-content" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.941731 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerName="extract-content" Dec 05 09:33:02 crc kubenswrapper[4876]: E1205 09:33:02.941744 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerName="registry-server" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.941752 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerName="registry-server" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.942052 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="51fa92df-ab98-43a6-9b77-ae5a85f84e36" containerName="registry-server" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.942064 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="de75f30c-84a7-49d8-9e16-1761c433d0b9" containerName="registry-server" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.943836 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.950771 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-79kf4"] Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.983188 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-utilities\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.983276 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq2sc\" (UniqueName: \"kubernetes.io/projected/e809f38f-3fc4-48ea-a7ad-2356a8378a39-kube-api-access-pq2sc\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:02 crc kubenswrapper[4876]: I1205 09:33:02.983367 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-catalog-content\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:03 crc kubenswrapper[4876]: I1205 09:33:03.085477 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-utilities\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:03 crc kubenswrapper[4876]: I1205 09:33:03.085633 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq2sc\" (UniqueName: \"kubernetes.io/projected/e809f38f-3fc4-48ea-a7ad-2356a8378a39-kube-api-access-pq2sc\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:03 crc kubenswrapper[4876]: I1205 09:33:03.085754 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-catalog-content\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:03 crc kubenswrapper[4876]: I1205 09:33:03.086621 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-catalog-content\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:03 crc kubenswrapper[4876]: I1205 09:33:03.087432 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-utilities\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:03 crc kubenswrapper[4876]: I1205 09:33:03.112287 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq2sc\" (UniqueName: \"kubernetes.io/projected/e809f38f-3fc4-48ea-a7ad-2356a8378a39-kube-api-access-pq2sc\") pod \"redhat-marketplace-79kf4\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:03 crc kubenswrapper[4876]: I1205 09:33:03.282568 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:03 crc kubenswrapper[4876]: I1205 09:33:03.756170 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-79kf4"] Dec 05 09:33:03 crc kubenswrapper[4876]: W1205 09:33:03.759576 4876 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode809f38f_3fc4_48ea_a7ad_2356a8378a39.slice/crio-0b76c57d3f795ddbb94066a0fd3b2aecb601ec8490e6c1cdf203abbf3adc1c86 WatchSource:0}: Error finding container 0b76c57d3f795ddbb94066a0fd3b2aecb601ec8490e6c1cdf203abbf3adc1c86: Status 404 returned error can't find the container with id 0b76c57d3f795ddbb94066a0fd3b2aecb601ec8490e6c1cdf203abbf3adc1c86 Dec 05 09:33:04 crc kubenswrapper[4876]: I1205 09:33:04.603178 4876 generic.go:334] "Generic (PLEG): container finished" podID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerID="a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a" exitCode=0 Dec 05 09:33:04 crc kubenswrapper[4876]: I1205 09:33:04.603240 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79kf4" event={"ID":"e809f38f-3fc4-48ea-a7ad-2356a8378a39","Type":"ContainerDied","Data":"a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a"} Dec 05 09:33:04 crc kubenswrapper[4876]: I1205 09:33:04.603523 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79kf4" event={"ID":"e809f38f-3fc4-48ea-a7ad-2356a8378a39","Type":"ContainerStarted","Data":"0b76c57d3f795ddbb94066a0fd3b2aecb601ec8490e6c1cdf203abbf3adc1c86"} Dec 05 09:33:06 crc kubenswrapper[4876]: I1205 09:33:06.627280 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79kf4" event={"ID":"e809f38f-3fc4-48ea-a7ad-2356a8378a39","Type":"ContainerStarted","Data":"2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8"} Dec 05 09:33:06 crc kubenswrapper[4876]: E1205 09:33:06.860298 4876 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode809f38f_3fc4_48ea_a7ad_2356a8378a39.slice/crio-2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode809f38f_3fc4_48ea_a7ad_2356a8378a39.slice/crio-conmon-2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:33:07 crc kubenswrapper[4876]: I1205 09:33:07.641366 4876 generic.go:334] "Generic (PLEG): container finished" podID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerID="2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8" exitCode=0 Dec 05 09:33:07 crc kubenswrapper[4876]: I1205 09:33:07.641440 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79kf4" event={"ID":"e809f38f-3fc4-48ea-a7ad-2356a8378a39","Type":"ContainerDied","Data":"2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8"} Dec 05 09:33:08 crc kubenswrapper[4876]: I1205 09:33:08.660229 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79kf4" event={"ID":"e809f38f-3fc4-48ea-a7ad-2356a8378a39","Type":"ContainerStarted","Data":"6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e"} Dec 05 09:33:08 crc kubenswrapper[4876]: I1205 09:33:08.683705 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-79kf4" podStartSLOduration=3.101247715 podStartE2EDuration="6.683683876s" podCreationTimestamp="2025-12-05 09:33:02 +0000 UTC" firstStartedPulling="2025-12-05 09:33:04.60497875 +0000 UTC m=+4649.093643372" lastFinishedPulling="2025-12-05 09:33:08.187414911 +0000 UTC m=+4652.676079533" observedRunningTime="2025-12-05 09:33:08.679986536 +0000 UTC m=+4653.168651158" watchObservedRunningTime="2025-12-05 09:33:08.683683876 +0000 UTC m=+4653.172348498" Dec 05 09:33:13 crc kubenswrapper[4876]: I1205 09:33:13.283188 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:13 crc kubenswrapper[4876]: I1205 09:33:13.286046 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:13 crc kubenswrapper[4876]: I1205 09:33:13.332599 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:13 crc kubenswrapper[4876]: I1205 09:33:13.774070 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:13 crc kubenswrapper[4876]: I1205 09:33:13.838804 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-79kf4"] Dec 05 09:33:15 crc kubenswrapper[4876]: I1205 09:33:15.725774 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-79kf4" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerName="registry-server" containerID="cri-o://6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e" gracePeriod=2 Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.165279 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.247292 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-catalog-content\") pod \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.247355 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-utilities\") pod \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.247419 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq2sc\" (UniqueName: \"kubernetes.io/projected/e809f38f-3fc4-48ea-a7ad-2356a8378a39-kube-api-access-pq2sc\") pod \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\" (UID: \"e809f38f-3fc4-48ea-a7ad-2356a8378a39\") " Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.249541 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-utilities" (OuterVolumeSpecName: "utilities") pod "e809f38f-3fc4-48ea-a7ad-2356a8378a39" (UID: "e809f38f-3fc4-48ea-a7ad-2356a8378a39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.255528 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e809f38f-3fc4-48ea-a7ad-2356a8378a39-kube-api-access-pq2sc" (OuterVolumeSpecName: "kube-api-access-pq2sc") pod "e809f38f-3fc4-48ea-a7ad-2356a8378a39" (UID: "e809f38f-3fc4-48ea-a7ad-2356a8378a39"). InnerVolumeSpecName "kube-api-access-pq2sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.271013 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e809f38f-3fc4-48ea-a7ad-2356a8378a39" (UID: "e809f38f-3fc4-48ea-a7ad-2356a8378a39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.349724 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq2sc\" (UniqueName: \"kubernetes.io/projected/e809f38f-3fc4-48ea-a7ad-2356a8378a39-kube-api-access-pq2sc\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.349761 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.349772 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e809f38f-3fc4-48ea-a7ad-2356a8378a39-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.735948 4876 generic.go:334] "Generic (PLEG): container finished" podID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerID="6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e" exitCode=0 Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.736063 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79kf4" event={"ID":"e809f38f-3fc4-48ea-a7ad-2356a8378a39","Type":"ContainerDied","Data":"6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e"} Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.737286 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-79kf4" event={"ID":"e809f38f-3fc4-48ea-a7ad-2356a8378a39","Type":"ContainerDied","Data":"0b76c57d3f795ddbb94066a0fd3b2aecb601ec8490e6c1cdf203abbf3adc1c86"} Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.736128 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-79kf4" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.737367 4876 scope.go:117] "RemoveContainer" containerID="6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.766974 4876 scope.go:117] "RemoveContainer" containerID="2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8" Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.777702 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-79kf4"] Dec 05 09:33:16 crc kubenswrapper[4876]: I1205 09:33:16.789889 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-79kf4"] Dec 05 09:33:17 crc kubenswrapper[4876]: I1205 09:33:17.381399 4876 scope.go:117] "RemoveContainer" containerID="a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a" Dec 05 09:33:17 crc kubenswrapper[4876]: I1205 09:33:17.431660 4876 scope.go:117] "RemoveContainer" containerID="6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e" Dec 05 09:33:17 crc kubenswrapper[4876]: E1205 09:33:17.432569 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e\": container with ID starting with 6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e not found: ID does not exist" containerID="6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e" Dec 05 09:33:17 crc kubenswrapper[4876]: I1205 09:33:17.432599 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e"} err="failed to get container status \"6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e\": rpc error: code = NotFound desc = could not find container \"6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e\": container with ID starting with 6adc0cfc9bf751930c626c7bdda495cb5f5e7fa8488e295cc472520753e1878e not found: ID does not exist" Dec 05 09:33:17 crc kubenswrapper[4876]: I1205 09:33:17.432616 4876 scope.go:117] "RemoveContainer" containerID="2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8" Dec 05 09:33:17 crc kubenswrapper[4876]: E1205 09:33:17.433028 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8\": container with ID starting with 2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8 not found: ID does not exist" containerID="2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8" Dec 05 09:33:17 crc kubenswrapper[4876]: I1205 09:33:17.433075 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8"} err="failed to get container status \"2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8\": rpc error: code = NotFound desc = could not find container \"2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8\": container with ID starting with 2f2fafaecb6af77263a4a5f17fa9528e11f646ed0f6b0ce87912d565a95cddc8 not found: ID does not exist" Dec 05 09:33:17 crc kubenswrapper[4876]: I1205 09:33:17.433101 4876 scope.go:117] "RemoveContainer" containerID="a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a" Dec 05 09:33:17 crc kubenswrapper[4876]: E1205 09:33:17.433511 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a\": container with ID starting with a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a not found: ID does not exist" containerID="a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a" Dec 05 09:33:17 crc kubenswrapper[4876]: I1205 09:33:17.433544 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a"} err="failed to get container status \"a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a\": rpc error: code = NotFound desc = could not find container \"a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a\": container with ID starting with a7cf52339d5766795f644d51665b43ba6e1f5433ac2ae0a94f3da65b3971a43a not found: ID does not exist" Dec 05 09:33:17 crc kubenswrapper[4876]: I1205 09:33:17.837180 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" path="/var/lib/kubelet/pods/e809f38f-3fc4-48ea-a7ad-2356a8378a39/volumes" Dec 05 09:33:47 crc kubenswrapper[4876]: I1205 09:33:47.067217 4876 generic.go:334] "Generic (PLEG): container finished" podID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerID="e6ad7bce468b0e870c0afdae7ef3e84fa59faf2189f4983772ebeb68f825908e" exitCode=0 Dec 05 09:33:47 crc kubenswrapper[4876]: I1205 09:33:47.067286 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" event={"ID":"4a5a3968-3ce7-4022-8687-c9de4995e1fd","Type":"ContainerDied","Data":"e6ad7bce468b0e870c0afdae7ef3e84fa59faf2189f4983772ebeb68f825908e"} Dec 05 09:33:47 crc kubenswrapper[4876]: I1205 09:33:47.069869 4876 scope.go:117] "RemoveContainer" containerID="e6ad7bce468b0e870c0afdae7ef3e84fa59faf2189f4983772ebeb68f825908e" Dec 05 09:33:47 crc kubenswrapper[4876]: I1205 09:33:47.175514 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-z2wdt_must-gather-cc4mq_4a5a3968-3ce7-4022-8687-c9de4995e1fd/gather/0.log" Dec 05 09:33:56 crc kubenswrapper[4876]: I1205 09:33:56.997709 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-z2wdt/must-gather-cc4mq"] Dec 05 09:33:56 crc kubenswrapper[4876]: I1205 09:33:56.999681 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" podUID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerName="copy" containerID="cri-o://882906be0595d14f56881d44a236e6b823622955bf329fa4c7c912439df0b67e" gracePeriod=2 Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.008453 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-z2wdt/must-gather-cc4mq"] Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.181094 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-z2wdt_must-gather-cc4mq_4a5a3968-3ce7-4022-8687-c9de4995e1fd/copy/0.log" Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.181822 4876 generic.go:334] "Generic (PLEG): container finished" podID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerID="882906be0595d14f56881d44a236e6b823622955bf329fa4c7c912439df0b67e" exitCode=143 Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.433707 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-z2wdt_must-gather-cc4mq_4a5a3968-3ce7-4022-8687-c9de4995e1fd/copy/0.log" Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.434819 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.507227 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4a5a3968-3ce7-4022-8687-c9de4995e1fd-must-gather-output\") pod \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\" (UID: \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\") " Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.507534 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljz45\" (UniqueName: \"kubernetes.io/projected/4a5a3968-3ce7-4022-8687-c9de4995e1fd-kube-api-access-ljz45\") pod \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\" (UID: \"4a5a3968-3ce7-4022-8687-c9de4995e1fd\") " Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.515159 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a5a3968-3ce7-4022-8687-c9de4995e1fd-kube-api-access-ljz45" (OuterVolumeSpecName: "kube-api-access-ljz45") pod "4a5a3968-3ce7-4022-8687-c9de4995e1fd" (UID: "4a5a3968-3ce7-4022-8687-c9de4995e1fd"). InnerVolumeSpecName "kube-api-access-ljz45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.610236 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljz45\" (UniqueName: \"kubernetes.io/projected/4a5a3968-3ce7-4022-8687-c9de4995e1fd-kube-api-access-ljz45\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.663994 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a5a3968-3ce7-4022-8687-c9de4995e1fd-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "4a5a3968-3ce7-4022-8687-c9de4995e1fd" (UID: "4a5a3968-3ce7-4022-8687-c9de4995e1fd"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.711727 4876 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4a5a3968-3ce7-4022-8687-c9de4995e1fd-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:57 crc kubenswrapper[4876]: I1205 09:33:57.837133 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" path="/var/lib/kubelet/pods/4a5a3968-3ce7-4022-8687-c9de4995e1fd/volumes" Dec 05 09:33:58 crc kubenswrapper[4876]: I1205 09:33:58.197444 4876 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-z2wdt_must-gather-cc4mq_4a5a3968-3ce7-4022-8687-c9de4995e1fd/copy/0.log" Dec 05 09:33:58 crc kubenswrapper[4876]: I1205 09:33:58.205769 4876 scope.go:117] "RemoveContainer" containerID="882906be0595d14f56881d44a236e6b823622955bf329fa4c7c912439df0b67e" Dec 05 09:33:58 crc kubenswrapper[4876]: I1205 09:33:58.205841 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z2wdt/must-gather-cc4mq" Dec 05 09:33:58 crc kubenswrapper[4876]: I1205 09:33:58.266095 4876 scope.go:117] "RemoveContainer" containerID="e6ad7bce468b0e870c0afdae7ef3e84fa59faf2189f4983772ebeb68f825908e" Dec 05 09:34:44 crc kubenswrapper[4876]: I1205 09:34:44.069511 4876 scope.go:117] "RemoveContainer" containerID="2d4276f6753b12ac4a50cced02f42741adb5f507a3ba72f712fefd9b7d85a21e" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.678651 4876 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xk5fl"] Dec 05 09:35:03 crc kubenswrapper[4876]: E1205 09:35:03.679947 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerName="registry-server" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.679971 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerName="registry-server" Dec 05 09:35:03 crc kubenswrapper[4876]: E1205 09:35:03.680004 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerName="extract-utilities" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.680017 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerName="extract-utilities" Dec 05 09:35:03 crc kubenswrapper[4876]: E1205 09:35:03.680040 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerName="extract-content" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.680051 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerName="extract-content" Dec 05 09:35:03 crc kubenswrapper[4876]: E1205 09:35:03.680073 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerName="copy" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.680084 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerName="copy" Dec 05 09:35:03 crc kubenswrapper[4876]: E1205 09:35:03.680122 4876 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerName="gather" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.680135 4876 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerName="gather" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.680446 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerName="copy" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.680491 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="e809f38f-3fc4-48ea-a7ad-2356a8378a39" containerName="registry-server" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.680525 4876 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a5a3968-3ce7-4022-8687-c9de4995e1fd" containerName="gather" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.683417 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.693867 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xk5fl"] Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.698592 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-catalog-content\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.698718 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xz2z\" (UniqueName: \"kubernetes.io/projected/8c890403-5726-4d0e-a0e2-734b8eb4bace-kube-api-access-8xz2z\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.698842 4876 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-utilities\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.800308 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-catalog-content\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.800387 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xz2z\" (UniqueName: \"kubernetes.io/projected/8c890403-5726-4d0e-a0e2-734b8eb4bace-kube-api-access-8xz2z\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.800446 4876 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-utilities\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.801046 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-utilities\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.801233 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-catalog-content\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:03 crc kubenswrapper[4876]: I1205 09:35:03.822812 4876 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xz2z\" (UniqueName: \"kubernetes.io/projected/8c890403-5726-4d0e-a0e2-734b8eb4bace-kube-api-access-8xz2z\") pod \"certified-operators-xk5fl\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:04 crc kubenswrapper[4876]: I1205 09:35:04.005030 4876 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:04 crc kubenswrapper[4876]: I1205 09:35:04.571719 4876 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xk5fl"] Dec 05 09:35:04 crc kubenswrapper[4876]: I1205 09:35:04.818988 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xk5fl" event={"ID":"8c890403-5726-4d0e-a0e2-734b8eb4bace","Type":"ContainerStarted","Data":"451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b"} Dec 05 09:35:04 crc kubenswrapper[4876]: I1205 09:35:04.819355 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xk5fl" event={"ID":"8c890403-5726-4d0e-a0e2-734b8eb4bace","Type":"ContainerStarted","Data":"76f895b6e8be195a86f84950f2e0fe8f4457503caa27b064cc83287ec4a2accb"} Dec 05 09:35:05 crc kubenswrapper[4876]: I1205 09:35:05.863105 4876 generic.go:334] "Generic (PLEG): container finished" podID="8c890403-5726-4d0e-a0e2-734b8eb4bace" containerID="451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b" exitCode=0 Dec 05 09:35:05 crc kubenswrapper[4876]: I1205 09:35:05.863381 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xk5fl" event={"ID":"8c890403-5726-4d0e-a0e2-734b8eb4bace","Type":"ContainerDied","Data":"451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b"} Dec 05 09:35:05 crc kubenswrapper[4876]: I1205 09:35:05.863406 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xk5fl" event={"ID":"8c890403-5726-4d0e-a0e2-734b8eb4bace","Type":"ContainerStarted","Data":"5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3"} Dec 05 09:35:06 crc kubenswrapper[4876]: I1205 09:35:06.873099 4876 generic.go:334] "Generic (PLEG): container finished" podID="8c890403-5726-4d0e-a0e2-734b8eb4bace" containerID="5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3" exitCode=0 Dec 05 09:35:06 crc kubenswrapper[4876]: I1205 09:35:06.873132 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xk5fl" event={"ID":"8c890403-5726-4d0e-a0e2-734b8eb4bace","Type":"ContainerDied","Data":"5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3"} Dec 05 09:35:07 crc kubenswrapper[4876]: I1205 09:35:07.886995 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xk5fl" event={"ID":"8c890403-5726-4d0e-a0e2-734b8eb4bace","Type":"ContainerStarted","Data":"757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d"} Dec 05 09:35:07 crc kubenswrapper[4876]: I1205 09:35:07.912238 4876 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xk5fl" podStartSLOduration=2.444540051 podStartE2EDuration="4.912218991s" podCreationTimestamp="2025-12-05 09:35:03 +0000 UTC" firstStartedPulling="2025-12-05 09:35:04.820738562 +0000 UTC m=+4769.309403204" lastFinishedPulling="2025-12-05 09:35:07.288417522 +0000 UTC m=+4771.777082144" observedRunningTime="2025-12-05 09:35:07.90921718 +0000 UTC m=+4772.397881802" watchObservedRunningTime="2025-12-05 09:35:07.912218991 +0000 UTC m=+4772.400883623" Dec 05 09:35:08 crc kubenswrapper[4876]: I1205 09:35:08.213612 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:35:08 crc kubenswrapper[4876]: I1205 09:35:08.213710 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:35:14 crc kubenswrapper[4876]: I1205 09:35:14.005954 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:14 crc kubenswrapper[4876]: I1205 09:35:14.007494 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:14 crc kubenswrapper[4876]: I1205 09:35:14.303248 4876 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:15 crc kubenswrapper[4876]: I1205 09:35:15.003868 4876 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:15 crc kubenswrapper[4876]: I1205 09:35:15.056246 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xk5fl"] Dec 05 09:35:16 crc kubenswrapper[4876]: I1205 09:35:16.978843 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xk5fl" podUID="8c890403-5726-4d0e-a0e2-734b8eb4bace" containerName="registry-server" containerID="cri-o://757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d" gracePeriod=2 Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.587239 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.771556 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-utilities\") pod \"8c890403-5726-4d0e-a0e2-734b8eb4bace\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.771736 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-catalog-content\") pod \"8c890403-5726-4d0e-a0e2-734b8eb4bace\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.771831 4876 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xz2z\" (UniqueName: \"kubernetes.io/projected/8c890403-5726-4d0e-a0e2-734b8eb4bace-kube-api-access-8xz2z\") pod \"8c890403-5726-4d0e-a0e2-734b8eb4bace\" (UID: \"8c890403-5726-4d0e-a0e2-734b8eb4bace\") " Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.772617 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-utilities" (OuterVolumeSpecName: "utilities") pod "8c890403-5726-4d0e-a0e2-734b8eb4bace" (UID: "8c890403-5726-4d0e-a0e2-734b8eb4bace"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.777120 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c890403-5726-4d0e-a0e2-734b8eb4bace-kube-api-access-8xz2z" (OuterVolumeSpecName: "kube-api-access-8xz2z") pod "8c890403-5726-4d0e-a0e2-734b8eb4bace" (UID: "8c890403-5726-4d0e-a0e2-734b8eb4bace"). InnerVolumeSpecName "kube-api-access-8xz2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.829194 4876 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c890403-5726-4d0e-a0e2-734b8eb4bace" (UID: "8c890403-5726-4d0e-a0e2-734b8eb4bace"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.875644 4876 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xz2z\" (UniqueName: \"kubernetes.io/projected/8c890403-5726-4d0e-a0e2-734b8eb4bace-kube-api-access-8xz2z\") on node \"crc\" DevicePath \"\"" Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.876009 4876 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.876094 4876 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c890403-5726-4d0e-a0e2-734b8eb4bace-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.988834 4876 generic.go:334] "Generic (PLEG): container finished" podID="8c890403-5726-4d0e-a0e2-734b8eb4bace" containerID="757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d" exitCode=0 Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.988888 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xk5fl" event={"ID":"8c890403-5726-4d0e-a0e2-734b8eb4bace","Type":"ContainerDied","Data":"757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d"} Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.988919 4876 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xk5fl" Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.988932 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xk5fl" event={"ID":"8c890403-5726-4d0e-a0e2-734b8eb4bace","Type":"ContainerDied","Data":"76f895b6e8be195a86f84950f2e0fe8f4457503caa27b064cc83287ec4a2accb"} Dec 05 09:35:17 crc kubenswrapper[4876]: I1205 09:35:17.988952 4876 scope.go:117] "RemoveContainer" containerID="757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d" Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.011499 4876 scope.go:117] "RemoveContainer" containerID="5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3" Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.019886 4876 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xk5fl"] Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.031097 4876 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xk5fl"] Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.040175 4876 scope.go:117] "RemoveContainer" containerID="451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b" Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.078941 4876 scope.go:117] "RemoveContainer" containerID="757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d" Dec 05 09:35:18 crc kubenswrapper[4876]: E1205 09:35:18.079380 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d\": container with ID starting with 757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d not found: ID does not exist" containerID="757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d" Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.079407 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d"} err="failed to get container status \"757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d\": rpc error: code = NotFound desc = could not find container \"757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d\": container with ID starting with 757bba523f32058c689245f3286f7b9799083694aa388f97294aea64af58665d not found: ID does not exist" Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.079439 4876 scope.go:117] "RemoveContainer" containerID="5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3" Dec 05 09:35:18 crc kubenswrapper[4876]: E1205 09:35:18.079798 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3\": container with ID starting with 5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3 not found: ID does not exist" containerID="5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3" Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.079853 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3"} err="failed to get container status \"5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3\": rpc error: code = NotFound desc = could not find container \"5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3\": container with ID starting with 5eae5116762a4b14bf0a6e03cb153975d0a16dd67057e00d0947ec295eeb41e3 not found: ID does not exist" Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.079884 4876 scope.go:117] "RemoveContainer" containerID="451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b" Dec 05 09:35:18 crc kubenswrapper[4876]: E1205 09:35:18.080326 4876 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b\": container with ID starting with 451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b not found: ID does not exist" containerID="451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b" Dec 05 09:35:18 crc kubenswrapper[4876]: I1205 09:35:18.080354 4876 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b"} err="failed to get container status \"451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b\": rpc error: code = NotFound desc = could not find container \"451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b\": container with ID starting with 451c26b89e2b3f45972401ee9162760d0a60ad30134e111caeb730c55d7daf1b not found: ID does not exist" Dec 05 09:35:19 crc kubenswrapper[4876]: I1205 09:35:19.836370 4876 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c890403-5726-4d0e-a0e2-734b8eb4bace" path="/var/lib/kubelet/pods/8c890403-5726-4d0e-a0e2-734b8eb4bace/volumes" Dec 05 09:35:38 crc kubenswrapper[4876]: I1205 09:35:38.214994 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:35:38 crc kubenswrapper[4876]: I1205 09:35:38.215534 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:36:08 crc kubenswrapper[4876]: I1205 09:36:08.213573 4876 patch_prober.go:28] interesting pod/machine-config-daemon-xj8qx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:36:08 crc kubenswrapper[4876]: I1205 09:36:08.214276 4876 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:36:08 crc kubenswrapper[4876]: I1205 09:36:08.214332 4876 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" Dec 05 09:36:08 crc kubenswrapper[4876]: I1205 09:36:08.216772 4876 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c754475cbb8158c1c8983f993d9812c384247e50a836b7438f9ae73893ac0eb"} pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:36:08 crc kubenswrapper[4876]: I1205 09:36:08.216858 4876 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" podUID="77322cc8-c6ab-4250-8098-9938309f0af8" containerName="machine-config-daemon" containerID="cri-o://5c754475cbb8158c1c8983f993d9812c384247e50a836b7438f9ae73893ac0eb" gracePeriod=600 Dec 05 09:36:08 crc kubenswrapper[4876]: I1205 09:36:08.444249 4876 generic.go:334] "Generic (PLEG): container finished" podID="77322cc8-c6ab-4250-8098-9938309f0af8" containerID="5c754475cbb8158c1c8983f993d9812c384247e50a836b7438f9ae73893ac0eb" exitCode=0 Dec 05 09:36:08 crc kubenswrapper[4876]: I1205 09:36:08.444328 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerDied","Data":"5c754475cbb8158c1c8983f993d9812c384247e50a836b7438f9ae73893ac0eb"} Dec 05 09:36:08 crc kubenswrapper[4876]: I1205 09:36:08.444490 4876 scope.go:117] "RemoveContainer" containerID="efdb313a77a9063c0cf124f4e69843fd0bb069170c35b9422ca3b6057fa13e87" Dec 05 09:36:09 crc kubenswrapper[4876]: I1205 09:36:09.456281 4876 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xj8qx" event={"ID":"77322cc8-c6ab-4250-8098-9938309f0af8","Type":"ContainerStarted","Data":"99bb3b500bf2d8771bb4db57afe4bf2e89d8acb2e6fd5aece52096f653589b1d"} Dec 05 09:36:31 crc kubenswrapper[4876]: I1205 09:36:31.760474 4876 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="761374dd-700e-42fa-81d4-d560f97643be" containerName="galera" probeResult="failure" output="command timed out" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114523743024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114523743017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114511744016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114511744015460 5ustar corecore